13118 1727204062.08901: starting run ansible-playbook [core 2.17.4] config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.12/site-packages/ansible ansible collection location = /tmp/collections-G1p executable location = /usr/local/bin/ansible-playbook python version = 3.12.5 (main, Aug 23 2024, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-2)] (/usr/bin/python3.12) jinja version = 3.1.4 libyaml = True No config file found; using defaults 13118 1727204062.09897: Added group all to inventory 13118 1727204062.09899: Added group ungrouped to inventory 13118 1727204062.09904: Group all now contains ungrouped 13118 1727204062.09907: Examining possible inventory source: /tmp/network-M6W/inventory-5vW.yml 13118 1727204062.31550: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/cache 13118 1727204062.31624: Loading CacheModule 'memory' from /usr/local/lib/python3.12/site-packages/ansible/plugins/cache/memory.py 13118 1727204062.31649: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory 13118 1727204062.31723: Loading InventoryModule 'host_list' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/host_list.py 13118 1727204062.31803: Loaded config def from plugin (inventory/script) 13118 1727204062.31805: Loading InventoryModule 'script' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/script.py 13118 1727204062.31854: Loading InventoryModule 'auto' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/auto.py 13118 1727204062.31946: Loaded config def from plugin (inventory/yaml) 13118 1727204062.31951: Loading InventoryModule 'yaml' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/yaml.py 13118 1727204062.32075: Loading InventoryModule 'ini' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/ini.py 13118 1727204062.32549: Loading InventoryModule 'toml' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/toml.py 13118 1727204062.32552: Attempting to use plugin host_list (/usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/host_list.py) 13118 1727204062.32555: Attempting to use plugin script (/usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/script.py) 13118 1727204062.32561: Attempting to use plugin auto (/usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/auto.py) 13118 1727204062.32568: Loading data from /tmp/network-M6W/inventory-5vW.yml 13118 1727204062.32676: /tmp/network-M6W/inventory-5vW.yml was not parsable by auto 13118 1727204062.32752: Attempting to use plugin yaml (/usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/yaml.py) 13118 1727204062.32794: Loading data from /tmp/network-M6W/inventory-5vW.yml 13118 1727204062.32898: group all already in inventory 13118 1727204062.32905: set inventory_file for managed-node1 13118 1727204062.32909: set inventory_dir for managed-node1 13118 1727204062.32910: Added host managed-node1 to inventory 13118 1727204062.32924: Added host managed-node1 to group all 13118 1727204062.32926: set ansible_host for managed-node1 13118 1727204062.32927: set ansible_ssh_extra_args for managed-node1 13118 1727204062.32931: set inventory_file for managed-node2 13118 1727204062.32939: set inventory_dir for managed-node2 13118 1727204062.32940: Added host managed-node2 to inventory 13118 1727204062.32942: Added host managed-node2 to group all 13118 1727204062.32943: set ansible_host for managed-node2 13118 1727204062.32944: set ansible_ssh_extra_args for managed-node2 13118 1727204062.32947: set inventory_file for managed-node3 13118 1727204062.32956: set inventory_dir for managed-node3 13118 1727204062.32958: Added host managed-node3 to inventory 13118 1727204062.32959: Added host managed-node3 to group all 13118 1727204062.32960: set ansible_host for managed-node3 13118 1727204062.32961: set ansible_ssh_extra_args for managed-node3 13118 1727204062.32966: Reconcile groups and hosts in inventory. 13118 1727204062.32970: Group ungrouped now contains managed-node1 13118 1727204062.32972: Group ungrouped now contains managed-node2 13118 1727204062.32973: Group ungrouped now contains managed-node3 13118 1727204062.33056: '/usr/local/lib/python3.12/site-packages/ansible/plugins/vars/__init__' skipped due to reserved name 13118 1727204062.33187: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments 13118 1727204062.35366: Loading ModuleDocFragment 'vars_plugin_staging' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/vars_plugin_staging.py 13118 1727204062.35402: Loaded config def from plugin (vars/host_group_vars) 13118 1727204062.35406: Loading VarsModule 'host_group_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/vars/host_group_vars.py (found_in_cache=False, class_only=True) 13118 1727204062.35415: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/vars 13118 1727204062.35425: Loading VarsModule 'host_group_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/vars/host_group_vars.py (found_in_cache=True, class_only=False) 13118 1727204062.35476: Loading CacheModule 'memory' from /usr/local/lib/python3.12/site-packages/ansible/plugins/cache/memory.py (found_in_cache=True, class_only=False) 13118 1727204062.36310: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204062.36414: Loading ModuleDocFragment 'connection_pipelining' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/connection_pipelining.py 13118 1727204062.36454: Loaded config def from plugin (connection/local) 13118 1727204062.36457: Loading Connection 'local' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/local.py (found_in_cache=False, class_only=True) 13118 1727204062.37169: Loaded config def from plugin (connection/paramiko_ssh) 13118 1727204062.37172: Loading Connection 'paramiko_ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/paramiko_ssh.py (found_in_cache=False, class_only=True) 13118 1727204062.38256: Loading ModuleDocFragment 'connection_pipelining' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/connection_pipelining.py (found_in_cache=True, class_only=False) 13118 1727204062.38307: Loaded config def from plugin (connection/psrp) 13118 1727204062.38310: Loading Connection 'psrp' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/psrp.py (found_in_cache=False, class_only=True) 13118 1727204062.39024: Loading ModuleDocFragment 'connection_pipelining' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/connection_pipelining.py (found_in_cache=True, class_only=False) 13118 1727204062.39075: Loaded config def from plugin (connection/ssh) 13118 1727204062.39078: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=False, class_only=True) 13118 1727204062.39458: Loading ModuleDocFragment 'connection_pipelining' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/connection_pipelining.py (found_in_cache=True, class_only=False) 13118 1727204062.39507: Loaded config def from plugin (connection/winrm) 13118 1727204062.39510: Loading Connection 'winrm' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/winrm.py (found_in_cache=False, class_only=True) 13118 1727204062.39542: '/usr/local/lib/python3.12/site-packages/ansible/plugins/shell/__init__' skipped due to reserved name 13118 1727204062.39617: Loading ModuleDocFragment 'shell_windows' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/shell_windows.py 13118 1727204062.39684: Loaded config def from plugin (shell/cmd) 13118 1727204062.39687: Loading ShellModule 'cmd' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/cmd.py (found_in_cache=False, class_only=True) 13118 1727204062.39726: Loading ModuleDocFragment 'shell_windows' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/shell_windows.py (found_in_cache=True, class_only=False) 13118 1727204062.39792: Loaded config def from plugin (shell/powershell) 13118 1727204062.39794: Loading ShellModule 'powershell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/powershell.py (found_in_cache=False, class_only=True) 13118 1727204062.39857: Loading ModuleDocFragment 'shell_common' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/shell_common.py 13118 1727204062.40052: Loaded config def from plugin (shell/sh) 13118 1727204062.40054: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=False, class_only=True) 13118 1727204062.40089: '/usr/local/lib/python3.12/site-packages/ansible/plugins/become/__init__' skipped due to reserved name 13118 1727204062.40274: Loaded config def from plugin (become/runas) 13118 1727204062.40277: Loading BecomeModule 'runas' from /usr/local/lib/python3.12/site-packages/ansible/plugins/become/runas.py (found_in_cache=False, class_only=True) 13118 1727204062.40512: Loaded config def from plugin (become/su) 13118 1727204062.40514: Loading BecomeModule 'su' from /usr/local/lib/python3.12/site-packages/ansible/plugins/become/su.py (found_in_cache=False, class_only=True) 13118 1727204062.40725: Loaded config def from plugin (become/sudo) 13118 1727204062.40727: Loading BecomeModule 'sudo' from /usr/local/lib/python3.12/site-packages/ansible/plugins/become/sudo.py (found_in_cache=False, class_only=True) running playbook inside collection fedora.linux_system_roles 13118 1727204062.40762: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tests_bond_nm.yml 13118 1727204062.41126: in VariableManager get_vars() 13118 1727204062.41211: done with get_vars() 13118 1727204062.41446: trying /usr/local/lib/python3.12/site-packages/ansible/modules 13118 1727204062.45694: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/action 13118 1727204062.45833: in VariableManager get_vars() 13118 1727204062.45839: done with get_vars() 13118 1727204062.45842: variable 'playbook_dir' from source: magic vars 13118 1727204062.45846: variable 'ansible_playbook_python' from source: magic vars 13118 1727204062.45847: variable 'ansible_config_file' from source: magic vars 13118 1727204062.45848: variable 'groups' from source: magic vars 13118 1727204062.45849: variable 'omit' from source: magic vars 13118 1727204062.45850: variable 'ansible_version' from source: magic vars 13118 1727204062.45851: variable 'ansible_check_mode' from source: magic vars 13118 1727204062.45851: variable 'ansible_diff_mode' from source: magic vars 13118 1727204062.45852: variable 'ansible_forks' from source: magic vars 13118 1727204062.45853: variable 'ansible_inventory_sources' from source: magic vars 13118 1727204062.45853: variable 'ansible_skip_tags' from source: magic vars 13118 1727204062.45857: variable 'ansible_limit' from source: magic vars 13118 1727204062.45858: variable 'ansible_run_tags' from source: magic vars 13118 1727204062.45859: variable 'ansible_verbosity' from source: magic vars 13118 1727204062.45939: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml 13118 1727204062.47016: in VariableManager get_vars() 13118 1727204062.47032: done with get_vars() 13118 1727204062.47041: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml statically imported: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml 13118 1727204062.48141: in VariableManager get_vars() 13118 1727204062.48158: done with get_vars() 13118 1727204062.48292: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml statically imported: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml 13118 1727204062.48502: in VariableManager get_vars() 13118 1727204062.48519: done with get_vars() 13118 1727204062.48680: in VariableManager get_vars() 13118 1727204062.48694: done with get_vars() 13118 1727204062.48703: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml statically imported: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml 13118 1727204062.48784: in VariableManager get_vars() 13118 1727204062.48799: done with get_vars() 13118 1727204062.49453: in VariableManager get_vars() 13118 1727204062.49471: done with get_vars() 13118 1727204062.49570: variable 'omit' from source: magic vars 13118 1727204062.49621: variable 'omit' from source: magic vars 13118 1727204062.49657: in VariableManager get_vars() 13118 1727204062.49671: done with get_vars() 13118 1727204062.49915: in VariableManager get_vars() 13118 1727204062.49929: done with get_vars() 13118 1727204062.49969: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/defaults/main.yml 13118 1727204062.50577: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/meta/main.yml 13118 1727204062.50982: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml 13118 1727204062.51858: in VariableManager get_vars() 13118 1727204062.51884: done with get_vars() 13118 1727204062.52473: trying /usr/local/lib/python3.12/site-packages/ansible/modules/__pycache__ 13118 1727204062.52695: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__ redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf 13118 1727204062.56475: in VariableManager get_vars() 13118 1727204062.56500: done with get_vars() 13118 1727204062.56511: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml statically imported: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml 13118 1727204062.56645: in VariableManager get_vars() 13118 1727204062.56665: done with get_vars() 13118 1727204062.56817: in VariableManager get_vars() 13118 1727204062.56835: done with get_vars() 13118 1727204062.57270: in VariableManager get_vars() 13118 1727204062.57290: done with get_vars() 13118 1727204062.57338: variable 'omit' from source: magic vars 13118 1727204062.57373: variable 'omit' from source: magic vars 13118 1727204062.57979: in VariableManager get_vars() 13118 1727204062.58001: done with get_vars() 13118 1727204062.58031: in VariableManager get_vars() 13118 1727204062.58048: done with get_vars() 13118 1727204062.58084: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/defaults/main.yml 13118 1727204062.58230: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/meta/main.yml 13118 1727204062.58312: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml 13118 1727204062.58886: in VariableManager get_vars() 13118 1727204062.58912: done with get_vars() redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf 13118 1727204062.61960: in VariableManager get_vars() 13118 1727204062.61993: done with get_vars() 13118 1727204062.62004: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml statically imported: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml 13118 1727204062.62557: in VariableManager get_vars() 13118 1727204062.62582: done with get_vars() 13118 1727204062.62647: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/callback 13118 1727204062.62661: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__ redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug 13118 1727204062.62911: Loading ModuleDocFragment 'default_callback' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/default_callback.py 13118 1727204062.63088: Loaded config def from plugin (callback/ansible_collections.ansible.posix.plugins.callback.debug) 13118 1727204062.63091: Loading CallbackModule 'ansible_collections.ansible.posix.plugins.callback.debug' from /tmp/collections-G1p/ansible_collections/ansible/posix/plugins/callback/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback) 13118 1727204062.63123: '/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__init__' skipped due to reserved name 13118 1727204062.63148: Loading ModuleDocFragment 'default_callback' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/default_callback.py (found_in_cache=True, class_only=False) 13118 1727204062.63338: Loading ModuleDocFragment 'result_format_callback' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/result_format_callback.py 13118 1727204062.63412: Loaded config def from plugin (callback/default) 13118 1727204062.63415: Loading CallbackModule 'default' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/default.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback) (found_in_cache=False, class_only=True) 13118 1727204062.69400: Loaded config def from plugin (callback/junit) 13118 1727204062.69404: Loading CallbackModule 'junit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/junit.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback) (found_in_cache=False, class_only=True) 13118 1727204062.69463: Loading ModuleDocFragment 'result_format_callback' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/result_format_callback.py (found_in_cache=True, class_only=False) 13118 1727204062.69542: Loaded config def from plugin (callback/minimal) 13118 1727204062.69544: Loading CallbackModule 'minimal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/minimal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback) (found_in_cache=False, class_only=True) 13118 1727204062.69591: Loading CallbackModule 'oneline' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/oneline.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback) (found_in_cache=False, class_only=True) 13118 1727204062.69666: Loaded config def from plugin (callback/tree) 13118 1727204062.69670: Loading CallbackModule 'tree' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/tree.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback) (found_in_cache=False, class_only=True) redirecting (type: callback) ansible.builtin.profile_tasks to ansible.posix.profile_tasks 13118 1727204062.70116: Loaded config def from plugin (callback/ansible_collections.ansible.posix.plugins.callback.profile_tasks) 13118 1727204062.70119: Loading CallbackModule 'ansible_collections.ansible.posix.plugins.callback.profile_tasks' from /tmp/collections-G1p/ansible_collections/ansible/posix/plugins/callback/profile_tasks.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback) (found_in_cache=False, class_only=True) Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_bond_nm.yml **************************************************** 2 plays in /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tests_bond_nm.yml 13118 1727204062.70151: in VariableManager get_vars() 13118 1727204062.70178: done with get_vars() 13118 1727204062.70188: in VariableManager get_vars() 13118 1727204062.70199: done with get_vars() 13118 1727204062.70203: variable 'omit' from source: magic vars 13118 1727204062.70246: in VariableManager get_vars() 13118 1727204062.70261: done with get_vars() 13118 1727204062.70309: variable 'omit' from source: magic vars PLAY [Run playbook 'playbooks/tests_bond.yml' with nm as provider] ************* 13118 1727204062.72122: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/strategy 13118 1727204062.72549: Loading StrategyModule 'linear' from /usr/local/lib/python3.12/site-packages/ansible/plugins/strategy/linear.py 13118 1727204062.72908: getting the remaining hosts for this loop 13118 1727204062.72910: done getting the remaining hosts for this loop 13118 1727204062.72914: getting the next task for host managed-node2 13118 1727204062.72917: done getting next task for host managed-node2 13118 1727204062.72919: ^ task is: TASK: Gathering Facts 13118 1727204062.72921: ^ state is: HOST STATE: block=0, task=0, rescue=0, always=0, handlers=0, run_state=0, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=True, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204062.72923: getting variables 13118 1727204062.72924: in VariableManager get_vars() 13118 1727204062.72939: Calling all_inventory to load vars for managed-node2 13118 1727204062.72941: Calling groups_inventory to load vars for managed-node2 13118 1727204062.72944: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204062.72959: Calling all_plugins_play to load vars for managed-node2 13118 1727204062.72972: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204062.72975: Calling groups_plugins_play to load vars for managed-node2 13118 1727204062.73010: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204062.73070: done with get_vars() 13118 1727204062.73078: done getting variables 13118 1727204062.73191: Loading ActionModule 'gather_facts' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/gather_facts.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=False, class_only=True) TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tests_bond_nm.yml:6 Tuesday 24 September 2024 14:54:22 -0400 (0:00:00.032) 0:00:00.032 ***** 13118 1727204062.73216: entering _queue_task() for managed-node2/gather_facts 13118 1727204062.73218: Creating lock for gather_facts 13118 1727204062.74215: worker is 1 (out of 1 available) 13118 1727204062.74230: exiting _queue_task() for managed-node2/gather_facts 13118 1727204062.74246: done queuing things up, now waiting for results queue to drain 13118 1727204062.74249: waiting for pending results... 13118 1727204062.74741: running TaskExecutor() for managed-node2/TASK: Gathering Facts 13118 1727204062.74851: in run() - task 0affcd87-79f5-56a3-0a64-0000000000cc 13118 1727204062.74882: variable 'ansible_search_path' from source: unknown 13118 1727204062.74920: calling self._execute() 13118 1727204062.74994: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204062.75006: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204062.75020: variable 'omit' from source: magic vars 13118 1727204062.75134: variable 'omit' from source: magic vars 13118 1727204062.75169: variable 'omit' from source: magic vars 13118 1727204062.75211: variable 'omit' from source: magic vars 13118 1727204062.75257: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204062.75302: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204062.75365: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204062.75435: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204062.75483: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204062.75519: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204062.75581: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204062.75589: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204062.76071: Set connection var ansible_timeout to 10 13118 1727204062.76098: Set connection var ansible_pipelining to False 13118 1727204062.76107: Set connection var ansible_connection to ssh 13118 1727204062.76116: Set connection var ansible_shell_executable to /bin/sh 13118 1727204062.76359: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204062.76370: Set connection var ansible_shell_type to sh 13118 1727204062.76407: variable 'ansible_shell_executable' from source: unknown 13118 1727204062.76415: variable 'ansible_connection' from source: unknown 13118 1727204062.76423: variable 'ansible_module_compression' from source: unknown 13118 1727204062.76433: variable 'ansible_shell_type' from source: unknown 13118 1727204062.76441: variable 'ansible_shell_executable' from source: unknown 13118 1727204062.76448: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204062.76456: variable 'ansible_pipelining' from source: unknown 13118 1727204062.76462: variable 'ansible_timeout' from source: unknown 13118 1727204062.76473: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204062.76680: Loading ActionModule 'gather_facts' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/gather_facts.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204062.76697: variable 'omit' from source: magic vars 13118 1727204062.76713: starting attempt loop 13118 1727204062.76721: running the handler 13118 1727204062.76745: variable 'ansible_facts' from source: unknown 13118 1727204062.77519: _low_level_execute_command(): starting 13118 1727204062.77539: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204062.78801: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204062.78822: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204062.78843: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204062.78872: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204062.78918: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204062.78935: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204062.78951: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204062.78980: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204062.78992: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204062.79001: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204062.79013: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204062.79025: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204062.79041: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204062.79051: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204062.79060: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204062.79077: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204062.79156: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204062.79189: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204062.79209: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204062.79288: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204062.80976: stdout chunk (state=3): >>>/root <<< 13118 1727204062.81182: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204062.81187: stdout chunk (state=3): >>><<< 13118 1727204062.81189: stderr chunk (state=3): >>><<< 13118 1727204062.81331: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204062.81335: _low_level_execute_command(): starting 13118 1727204062.81338: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204062.8121483-13211-79342327744476 `" && echo ansible-tmp-1727204062.8121483-13211-79342327744476="` echo /root/.ansible/tmp/ansible-tmp-1727204062.8121483-13211-79342327744476 `" ) && sleep 0' 13118 1727204062.82126: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204062.82307: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204062.82331: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204062.82352: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204062.82401: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204062.82418: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204062.82434: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204062.82451: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204062.82463: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204062.82478: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204062.82491: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204062.82505: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204062.82520: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204062.82536: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204062.82549: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204062.82562: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204062.82640: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204062.82660: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204062.82682: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204062.82861: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204062.84684: stdout chunk (state=3): >>>ansible-tmp-1727204062.8121483-13211-79342327744476=/root/.ansible/tmp/ansible-tmp-1727204062.8121483-13211-79342327744476 <<< 13118 1727204062.84871: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204062.84887: stdout chunk (state=3): >>><<< 13118 1727204062.84893: stderr chunk (state=3): >>><<< 13118 1727204062.84907: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204062.8121483-13211-79342327744476=/root/.ansible/tmp/ansible-tmp-1727204062.8121483-13211-79342327744476 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204062.84940: variable 'ansible_module_compression' from source: unknown 13118 1727204062.84996: ANSIBALLZ: Using generic lock for ansible.legacy.setup 13118 1727204062.84999: ANSIBALLZ: Acquiring lock 13118 1727204062.85002: ANSIBALLZ: Lock acquired: 140051944306976 13118 1727204062.85005: ANSIBALLZ: Creating module 13118 1727204063.20944: ANSIBALLZ: Writing module into payload 13118 1727204063.21245: ANSIBALLZ: Writing module 13118 1727204063.21286: ANSIBALLZ: Renaming module 13118 1727204063.21299: ANSIBALLZ: Done creating module 13118 1727204063.21342: variable 'ansible_facts' from source: unknown 13118 1727204063.21353: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204063.21393: _low_level_execute_command(): starting 13118 1727204063.21404: _low_level_execute_command(): executing: /bin/sh -c 'echo PLATFORM; uname; echo FOUND; command -v '"'"'python3.12'"'"'; command -v '"'"'python3.11'"'"'; command -v '"'"'python3.10'"'"'; command -v '"'"'python3.9'"'"'; command -v '"'"'python3.8'"'"'; command -v '"'"'python3.7'"'"'; command -v '"'"'/usr/bin/python3'"'"'; command -v '"'"'python3'"'"'; echo ENDFOUND && sleep 0' 13118 1727204063.22122: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204063.22145: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204063.22162: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204063.22187: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204063.22235: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204063.22251: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204063.22269: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204063.22289: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204063.22302: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204063.22314: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204063.22325: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204063.22343: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204063.22363: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204063.22378: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204063.22390: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204063.22404: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204063.22488: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204063.22506: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204063.22521: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204063.22606: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204063.24289: stdout chunk (state=3): >>>PLATFORM <<< 13118 1727204063.24390: stdout chunk (state=3): >>>Linux FOUND /usr/bin/python3.9 /usr/bin/python3 /usr/bin/python3 ENDFOUND <<< 13118 1727204063.24533: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204063.24630: stderr chunk (state=3): >>><<< 13118 1727204063.24645: stdout chunk (state=3): >>><<< 13118 1727204063.24779: _low_level_execute_command() done: rc=0, stdout=PLATFORM Linux FOUND /usr/bin/python3.9 /usr/bin/python3 /usr/bin/python3 ENDFOUND , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204063.24791 [managed-node2]: found interpreters: ['/usr/bin/python3.9', '/usr/bin/python3', '/usr/bin/python3'] 13118 1727204063.24794: _low_level_execute_command(): starting 13118 1727204063.24797: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 && sleep 0' 13118 1727204063.25290: Sending initial data 13118 1727204063.25294: Sent initial data (1181 bytes) 13118 1727204063.27054: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204063.27192: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204063.27254: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204063.27284: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204063.27349: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204063.27401: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204063.27436: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204063.27476: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204063.27498: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204063.27508: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204063.27518: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204063.27537: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204063.27550: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204063.27559: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204063.27569: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204063.27581: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204063.27670: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204063.27701: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204063.27719: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204063.27790: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204063.31597: stdout chunk (state=3): >>>{"platform_dist_result": [], "osrelease_content": "NAME=\"CentOS Stream\"\nVERSION=\"9\"\nID=\"centos\"\nID_LIKE=\"rhel fedora\"\nVERSION_ID=\"9\"\nPLATFORM_ID=\"platform:el9\"\nPRETTY_NAME=\"CentOS Stream 9\"\nANSI_COLOR=\"0;31\"\nLOGO=\"fedora-logo-icon\"\nCPE_NAME=\"cpe:/o:centos:centos:9\"\nHOME_URL=\"https://centos.org/\"\nBUG_REPORT_URL=\"https://issues.redhat.com/\"\nREDHAT_SUPPORT_PRODUCT=\"Red Hat Enterprise Linux 9\"\nREDHAT_SUPPORT_PRODUCT_VERSION=\"CentOS Stream\"\n"} <<< 13118 1727204063.32072: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204063.32076: stdout chunk (state=3): >>><<< 13118 1727204063.32274: stderr chunk (state=3): >>><<< 13118 1727204063.32278: _low_level_execute_command() done: rc=0, stdout={"platform_dist_result": [], "osrelease_content": "NAME=\"CentOS Stream\"\nVERSION=\"9\"\nID=\"centos\"\nID_LIKE=\"rhel fedora\"\nVERSION_ID=\"9\"\nPLATFORM_ID=\"platform:el9\"\nPRETTY_NAME=\"CentOS Stream 9\"\nANSI_COLOR=\"0;31\"\nLOGO=\"fedora-logo-icon\"\nCPE_NAME=\"cpe:/o:centos:centos:9\"\nHOME_URL=\"https://centos.org/\"\nBUG_REPORT_URL=\"https://issues.redhat.com/\"\nREDHAT_SUPPORT_PRODUCT=\"Red Hat Enterprise Linux 9\"\nREDHAT_SUPPORT_PRODUCT_VERSION=\"CentOS Stream\"\n"} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204063.32281: variable 'ansible_facts' from source: unknown 13118 1727204063.32284: variable 'ansible_facts' from source: unknown 13118 1727204063.32286: variable 'ansible_module_compression' from source: unknown 13118 1727204063.32288: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.setup-ZIP_DEFLATED 13118 1727204063.32371: variable 'ansible_facts' from source: unknown 13118 1727204063.32462: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204062.8121483-13211-79342327744476/AnsiballZ_setup.py 13118 1727204063.32938: Sending initial data 13118 1727204063.32947: Sent initial data (153 bytes) 13118 1727204063.35757: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204063.35780: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204063.35787: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204063.35801: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204063.35878: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204063.35882: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204063.35885: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204063.35887: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204063.35890: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204063.35892: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204063.36072: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204063.36075: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204063.36077: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204063.36079: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204063.36084: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204063.36086: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204063.36088: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204063.36090: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204063.36092: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204063.36171: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204063.37946: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204063.38026: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204063.38056: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204063.38982: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpiikfsp8w /root/.ansible/tmp/ansible-tmp-1727204062.8121483-13211-79342327744476/AnsiballZ_setup.py <<< 13118 1727204063.41191: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204063.41430: stderr chunk (state=3): >>><<< 13118 1727204063.41434: stdout chunk (state=3): >>><<< 13118 1727204063.41437: done transferring module to remote 13118 1727204063.41439: _low_level_execute_command(): starting 13118 1727204063.41441: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204062.8121483-13211-79342327744476/ /root/.ansible/tmp/ansible-tmp-1727204062.8121483-13211-79342327744476/AnsiballZ_setup.py && sleep 0' 13118 1727204063.42105: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204063.42110: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204063.42143: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204063.42146: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204063.42149: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204063.42151: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204063.42211: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204063.42236: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204063.42255: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204063.42336: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204063.44192: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204063.44235: stderr chunk (state=3): >>><<< 13118 1727204063.44238: stdout chunk (state=3): >>><<< 13118 1727204063.44344: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204063.44348: _low_level_execute_command(): starting 13118 1727204063.44353: _low_level_execute_command(): executing: /bin/sh -c 'PYTHONVERBOSE=1 /usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204062.8121483-13211-79342327744476/AnsiballZ_setup.py && sleep 0' 13118 1727204063.45552: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204063.45556: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204063.45597: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204063.45600: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204063.45602: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204063.45675: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204063.45678: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204063.45680: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204063.45742: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204063.47733: stdout chunk (state=3): >>>import _frozen_importlib # frozen import _imp # builtin import '_thread' # <<< 13118 1727204063.47738: stdout chunk (state=3): >>>import '_warnings' # import '_weakref' # <<< 13118 1727204063.47787: stdout chunk (state=3): >>>import '_io' # import 'marshal' # <<< 13118 1727204063.47821: stdout chunk (state=3): >>>import 'posix' # <<< 13118 1727204063.47855: stdout chunk (state=3): >>>import '_frozen_importlib_external' # <<< 13118 1727204063.47859: stdout chunk (state=3): >>># installing zipimport hook <<< 13118 1727204063.47894: stdout chunk (state=3): >>>import 'time' # <<< 13118 1727204063.47898: stdout chunk (state=3): >>>import 'zipimport' # # installed zipimport hook <<< 13118 1727204063.47943: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/encodings/__init__.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204063.47971: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/codecs.cpython-39.pyc matches /usr/lib64/python3.9/codecs.py <<< 13118 1727204063.47987: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/codecs.cpython-39.pyc' import '_codecs' # <<< 13118 1727204063.48011: stdout chunk (state=3): >>>import 'codecs' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a843dc0> <<< 13118 1727204063.48042: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/aliases.cpython-39.pyc matches /usr/lib64/python3.9/encodings/aliases.py <<< 13118 1727204063.48065: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/encodings/__pycache__/aliases.cpython-39.pyc' import 'encodings.aliases' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5d83a0> import 'encodings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a843b20> <<< 13118 1727204063.48096: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/utf_8.cpython-39.pyc matches /usr/lib64/python3.9/encodings/utf_8.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/utf_8.cpython-39.pyc' <<< 13118 1727204063.48111: stdout chunk (state=3): >>>import 'encodings.utf_8' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a843ac0> <<< 13118 1727204063.48128: stdout chunk (state=3): >>>import '_signal' # <<< 13118 1727204063.48157: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/latin_1.cpython-39.pyc matches /usr/lib64/python3.9/encodings/latin_1.py <<< 13118 1727204063.48160: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/encodings/__pycache__/latin_1.cpython-39.pyc' import 'encodings.latin_1' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5d8490> <<< 13118 1727204063.48199: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/io.cpython-39.pyc matches /usr/lib64/python3.9/io.py # code object from '/usr/lib64/python3.9/__pycache__/io.cpython-39.pyc' <<< 13118 1727204063.48203: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/abc.py <<< 13118 1727204063.48221: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/abc.cpython-39.pyc' import '_abc' # <<< 13118 1727204063.48233: stdout chunk (state=3): >>>import 'abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5d8940> <<< 13118 1727204063.48245: stdout chunk (state=3): >>>import 'io' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5d8670> <<< 13118 1727204063.48282: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/site.cpython-39.pyc matches /usr/lib64/python3.9/site.py # code object from '/usr/lib64/python3.9/__pycache__/site.cpython-39.pyc' <<< 13118 1727204063.48311: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/os.cpython-39.pyc matches /usr/lib64/python3.9/os.py <<< 13118 1727204063.48330: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/os.cpython-39.pyc' <<< 13118 1727204063.48346: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/stat.cpython-39.pyc matches /usr/lib64/python3.9/stat.py <<< 13118 1727204063.48365: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/stat.cpython-39.pyc' <<< 13118 1727204063.48387: stdout chunk (state=3): >>>import '_stat' # import 'stat' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a58f190> <<< 13118 1727204063.48409: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_collections_abc.cpython-39.pyc matches /usr/lib64/python3.9/_collections_abc.py <<< 13118 1727204063.48430: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/_collections_abc.cpython-39.pyc' <<< 13118 1727204063.48496: stdout chunk (state=3): >>>import '_collections_abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a58f220> <<< 13118 1727204063.48534: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/posixpath.cpython-39.pyc matches /usr/lib64/python3.9/posixpath.py <<< 13118 1727204063.48537: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/posixpath.cpython-39.pyc' <<< 13118 1727204063.48558: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/genericpath.cpython-39.pyc matches /usr/lib64/python3.9/genericpath.py # code object from '/usr/lib64/python3.9/__pycache__/genericpath.cpython-39.pyc' import 'genericpath' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5b2850> import 'posixpath' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a58f940> <<< 13118 1727204063.48589: stdout chunk (state=3): >>>import 'os' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5f0880> <<< 13118 1727204063.48621: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_sitebuiltins.cpython-39.pyc matches /usr/lib64/python3.9/_sitebuiltins.py <<< 13118 1727204063.48624: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/_sitebuiltins.cpython-39.pyc' <<< 13118 1727204063.48627: stdout chunk (state=3): >>>import '_sitebuiltins' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a588d90> <<< 13118 1727204063.48670: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_bootlocale.cpython-39.pyc matches /usr/lib64/python3.9/_bootlocale.py # code object from '/usr/lib64/python3.9/__pycache__/_bootlocale.cpython-39.pyc' <<< 13118 1727204063.48685: stdout chunk (state=3): >>>import '_locale' # import '_bootlocale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5b2d90> <<< 13118 1727204063.48737: stdout chunk (state=3): >>>import 'site' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5d8970> <<< 13118 1727204063.48762: stdout chunk (state=3): >>>Python 3.9.19 (main, Aug 23 2024, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-2)] on linux Type "help", "copyright", "credits" or "license" for more information. <<< 13118 1727204063.49093: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/base64.cpython-39.pyc matches /usr/lib64/python3.9/base64.py <<< 13118 1727204063.49114: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/base64.cpython-39.pyc' <<< 13118 1727204063.49134: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/re.cpython-39.pyc matches /usr/lib64/python3.9/re.py <<< 13118 1727204063.49138: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/re.cpython-39.pyc' <<< 13118 1727204063.49151: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/enum.cpython-39.pyc matches /usr/lib64/python3.9/enum.py <<< 13118 1727204063.49177: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/enum.cpython-39.pyc' <<< 13118 1727204063.49195: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/types.cpython-39.pyc matches /usr/lib64/python3.9/types.py <<< 13118 1727204063.49200: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/types.cpython-39.pyc' <<< 13118 1727204063.49218: stdout chunk (state=3): >>>import 'types' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a52ef10> <<< 13118 1727204063.49255: stdout chunk (state=3): >>>import 'enum' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5340a0> <<< 13118 1727204063.49283: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/sre_compile.cpython-39.pyc matches /usr/lib64/python3.9/sre_compile.py <<< 13118 1727204063.49286: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/sre_compile.cpython-39.pyc' <<< 13118 1727204063.49304: stdout chunk (state=3): >>>import '_sre' # <<< 13118 1727204063.49336: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/sre_parse.cpython-39.pyc matches /usr/lib64/python3.9/sre_parse.py <<< 13118 1727204063.49339: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/sre_parse.cpython-39.pyc' <<< 13118 1727204063.49365: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/sre_constants.cpython-39.pyc matches /usr/lib64/python3.9/sre_constants.py <<< 13118 1727204063.49370: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/sre_constants.cpython-39.pyc' <<< 13118 1727204063.49386: stdout chunk (state=3): >>>import 'sre_constants' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5275b0> <<< 13118 1727204063.49403: stdout chunk (state=3): >>>import 'sre_parse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a52f6a0> import 'sre_compile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a52e3d0> <<< 13118 1727204063.49427: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/functools.cpython-39.pyc matches /usr/lib64/python3.9/functools.py <<< 13118 1727204063.49523: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/functools.cpython-39.pyc' # /usr/lib64/python3.9/collections/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/collections/__init__.py <<< 13118 1727204063.49550: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/collections/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204063.49569: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/heapq.cpython-39.pyc matches /usr/lib64/python3.9/heapq.py # code object from '/usr/lib64/python3.9/__pycache__/heapq.cpython-39.pyc' <<< 13118 1727204063.49607: stdout chunk (state=3): >>># extension module '_heapq' loaded from '/usr/lib64/python3.9/lib-dynload/_heapq.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204063.49611: stdout chunk (state=3): >>># extension module '_heapq' executed from '/usr/lib64/python3.9/lib-dynload/_heapq.cpython-39-x86_64-linux-gnu.so' import '_heapq' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a415e50> import 'heapq' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a415940> <<< 13118 1727204063.49627: stdout chunk (state=3): >>>import 'itertools' # <<< 13118 1727204063.49650: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/keyword.cpython-39.pyc matches /usr/lib64/python3.9/keyword.py <<< 13118 1727204063.49655: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/keyword.cpython-39.pyc' import 'keyword' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a415f40> <<< 13118 1727204063.49677: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/operator.cpython-39.pyc matches /usr/lib64/python3.9/operator.py <<< 13118 1727204063.49705: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/operator.cpython-39.pyc' <<< 13118 1727204063.49708: stdout chunk (state=3): >>>import '_operator' # <<< 13118 1727204063.49712: stdout chunk (state=3): >>>import 'operator' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a415d90> <<< 13118 1727204063.49733: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/reprlib.cpython-39.pyc matches /usr/lib64/python3.9/reprlib.py <<< 13118 1727204063.49737: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/reprlib.cpython-39.pyc' import 'reprlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a426100> <<< 13118 1727204063.49749: stdout chunk (state=3): >>>import '_collections' # <<< 13118 1727204063.49801: stdout chunk (state=3): >>>import 'collections' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a509dc0> <<< 13118 1727204063.49804: stdout chunk (state=3): >>>import '_functools' # <<< 13118 1727204063.49828: stdout chunk (state=3): >>>import 'functools' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5026a0> <<< 13118 1727204063.49894: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/copyreg.cpython-39.pyc matches /usr/lib64/python3.9/copyreg.py # code object from '/usr/lib64/python3.9/__pycache__/copyreg.cpython-39.pyc' import 'copyreg' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a515700> <<< 13118 1727204063.49897: stdout chunk (state=3): >>>import 're' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a535eb0> <<< 13118 1727204063.49909: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/struct.cpython-39.pyc matches /usr/lib64/python3.9/struct.py # code object from '/usr/lib64/python3.9/__pycache__/struct.cpython-39.pyc' <<< 13118 1727204063.49944: stdout chunk (state=3): >>># extension module '_struct' loaded from '/usr/lib64/python3.9/lib-dynload/_struct.cpython-39-x86_64-linux-gnu.so' # extension module '_struct' executed from '/usr/lib64/python3.9/lib-dynload/_struct.cpython-39-x86_64-linux-gnu.so' import '_struct' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a426d00> <<< 13118 1727204063.49948: stdout chunk (state=3): >>>import 'struct' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5092e0> <<< 13118 1727204063.49990: stdout chunk (state=3): >>># extension module 'binascii' loaded from '/usr/lib64/python3.9/lib-dynload/binascii.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204063.49995: stdout chunk (state=3): >>># extension module 'binascii' executed from '/usr/lib64/python3.9/lib-dynload/binascii.cpython-39-x86_64-linux-gnu.so' import 'binascii' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a515310> <<< 13118 1727204063.49998: stdout chunk (state=3): >>>import 'base64' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a53ba60> <<< 13118 1727204063.50021: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/runpy.cpython-39.pyc matches /usr/lib64/python3.9/runpy.py # code object from '/usr/lib64/python3.9/__pycache__/runpy.cpython-39.pyc' <<< 13118 1727204063.50041: stdout chunk (state=3): >>># /usr/lib64/python3.9/importlib/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/importlib/__init__.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204063.50062: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/warnings.cpython-39.pyc matches /usr/lib64/python3.9/warnings.py <<< 13118 1727204063.50067: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/warnings.cpython-39.pyc' <<< 13118 1727204063.50085: stdout chunk (state=3): >>>import 'warnings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a426ee0> import 'importlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a426e20> <<< 13118 1727204063.50139: stdout chunk (state=3): >>># /usr/lib64/python3.9/importlib/__pycache__/machinery.cpython-39.pyc matches /usr/lib64/python3.9/importlib/machinery.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/machinery.cpython-39.pyc' <<< 13118 1727204063.50142: stdout chunk (state=3): >>>import 'importlib.machinery' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a426d90> <<< 13118 1727204063.50144: stdout chunk (state=3): >>># /usr/lib64/python3.9/importlib/__pycache__/util.cpython-39.pyc matches /usr/lib64/python3.9/importlib/util.py <<< 13118 1727204063.50147: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/importlib/__pycache__/util.cpython-39.pyc' <<< 13118 1727204063.50169: stdout chunk (state=3): >>># /usr/lib64/python3.9/importlib/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/importlib/abc.py <<< 13118 1727204063.50172: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/importlib/__pycache__/abc.cpython-39.pyc' <<< 13118 1727204063.50191: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/typing.cpython-39.pyc matches /usr/lib64/python3.9/typing.py <<< 13118 1727204063.50259: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/typing.cpython-39.pyc' <<< 13118 1727204063.50263: stdout chunk (state=3): >>># /usr/lib64/python3.9/collections/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/collections/abc.py # code object from '/usr/lib64/python3.9/collections/__pycache__/abc.cpython-39.pyc' <<< 13118 1727204063.50280: stdout chunk (state=3): >>>import 'collections.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a3f9400> # /usr/lib64/python3.9/__pycache__/contextlib.cpython-39.pyc matches /usr/lib64/python3.9/contextlib.py <<< 13118 1727204063.50300: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/contextlib.cpython-39.pyc' <<< 13118 1727204063.50327: stdout chunk (state=3): >>>import 'contextlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a3f94f0> <<< 13118 1727204063.50453: stdout chunk (state=3): >>>import 'typing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a42ef70> <<< 13118 1727204063.50482: stdout chunk (state=3): >>>import 'importlib.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a428ac0> <<< 13118 1727204063.50493: stdout chunk (state=3): >>>import 'importlib.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a428490> <<< 13118 1727204063.50523: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/pkgutil.cpython-39.pyc matches /usr/lib64/python3.9/pkgutil.py <<< 13118 1727204063.50526: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/pkgutil.cpython-39.pyc' <<< 13118 1727204063.50557: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/weakref.cpython-39.pyc matches /usr/lib64/python3.9/weakref.py <<< 13118 1727204063.50572: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/weakref.cpython-39.pyc' <<< 13118 1727204063.50588: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_weakrefset.cpython-39.pyc matches /usr/lib64/python3.9/_weakrefset.py <<< 13118 1727204063.50599: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/_weakrefset.cpython-39.pyc' import '_weakrefset' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a32d250> <<< 13118 1727204063.50635: stdout chunk (state=3): >>>import 'weakref' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a3e4550> <<< 13118 1727204063.50689: stdout chunk (state=3): >>>import 'pkgutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a428f40> <<< 13118 1727204063.50692: stdout chunk (state=3): >>>import 'runpy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a53b0d0> <<< 13118 1727204063.50705: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/shutil.cpython-39.pyc matches /usr/lib64/python3.9/shutil.py <<< 13118 1727204063.50772: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/shutil.cpython-39.pyc' <<< 13118 1727204063.50808: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/fnmatch.cpython-39.pyc matches /usr/lib64/python3.9/fnmatch.py # code object from '/usr/lib64/python3.9/__pycache__/fnmatch.cpython-39.pyc' import 'fnmatch' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a33fb80> import 'errno' # # extension module 'zlib' loaded from '/usr/lib64/python3.9/lib-dynload/zlib.cpython-39-x86_64-linux-gnu.so' # extension module 'zlib' executed from '/usr/lib64/python3.9/lib-dynload/zlib.cpython-39-x86_64-linux-gnu.so' import 'zlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a33feb0> <<< 13118 1727204063.50823: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/bz2.cpython-39.pyc matches /usr/lib64/python3.9/bz2.py # code object from '/usr/lib64/python3.9/__pycache__/bz2.cpython-39.pyc' <<< 13118 1727204063.50857: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_compression.cpython-39.pyc matches /usr/lib64/python3.9/_compression.py # code object from '/usr/lib64/python3.9/__pycache__/_compression.cpython-39.pyc' <<< 13118 1727204063.50860: stdout chunk (state=3): >>>import '_compression' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a3507c0> <<< 13118 1727204063.50897: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/threading.cpython-39.pyc matches /usr/lib64/python3.9/threading.py <<< 13118 1727204063.50913: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/threading.cpython-39.pyc' <<< 13118 1727204063.50943: stdout chunk (state=3): >>>import 'threading' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a350d00> <<< 13118 1727204063.50987: stdout chunk (state=3): >>># extension module '_bz2' loaded from '/usr/lib64/python3.9/lib-dynload/_bz2.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204063.50991: stdout chunk (state=3): >>># extension module '_bz2' executed from '/usr/lib64/python3.9/lib-dynload/_bz2.cpython-39-x86_64-linux-gnu.so' import '_bz2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a2de430> import 'bz2' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a33ffa0> <<< 13118 1727204063.51014: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/lzma.cpython-39.pyc matches /usr/lib64/python3.9/lzma.py # code object from '/usr/lib64/python3.9/__pycache__/lzma.cpython-39.pyc' <<< 13118 1727204063.51061: stdout chunk (state=3): >>># extension module '_lzma' loaded from '/usr/lib64/python3.9/lib-dynload/_lzma.cpython-39-x86_64-linux-gnu.so' # extension module '_lzma' executed from '/usr/lib64/python3.9/lib-dynload/_lzma.cpython-39-x86_64-linux-gnu.so' import '_lzma' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a2ef310> <<< 13118 1727204063.51070: stdout chunk (state=3): >>>import 'lzma' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a350640> <<< 13118 1727204063.51086: stdout chunk (state=3): >>>import 'pwd' # <<< 13118 1727204063.51101: stdout chunk (state=3): >>># extension module 'grp' loaded from '/usr/lib64/python3.9/lib-dynload/grp.cpython-39-x86_64-linux-gnu.so' # extension module 'grp' executed from '/usr/lib64/python3.9/lib-dynload/grp.cpython-39-x86_64-linux-gnu.so' import 'grp' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a2ef3d0> <<< 13118 1727204063.51149: stdout chunk (state=3): >>>import 'shutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a426a60> <<< 13118 1727204063.51153: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/tempfile.cpython-39.pyc matches /usr/lib64/python3.9/tempfile.py <<< 13118 1727204063.51174: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/tempfile.cpython-39.pyc' <<< 13118 1727204063.51193: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/random.cpython-39.pyc matches /usr/lib64/python3.9/random.py <<< 13118 1727204063.51207: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/random.cpython-39.pyc' <<< 13118 1727204063.51245: stdout chunk (state=3): >>># extension module 'math' loaded from '/usr/lib64/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204063.51250: stdout chunk (state=3): >>># extension module 'math' executed from '/usr/lib64/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so' import 'math' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a30b730> <<< 13118 1727204063.51263: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/bisect.cpython-39.pyc matches /usr/lib64/python3.9/bisect.py # code object from '/usr/lib64/python3.9/__pycache__/bisect.cpython-39.pyc' <<< 13118 1727204063.51290: stdout chunk (state=3): >>># extension module '_bisect' loaded from '/usr/lib64/python3.9/lib-dynload/_bisect.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204063.51293: stdout chunk (state=3): >>># extension module '_bisect' executed from '/usr/lib64/python3.9/lib-dynload/_bisect.cpython-39-x86_64-linux-gnu.so' import '_bisect' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a30ba00> import 'bisect' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a30b7f0> <<< 13118 1727204063.51309: stdout chunk (state=3): >>># extension module '_random' loaded from '/usr/lib64/python3.9/lib-dynload/_random.cpython-39-x86_64-linux-gnu.so' # extension module '_random' executed from '/usr/lib64/python3.9/lib-dynload/_random.cpython-39-x86_64-linux-gnu.so' import '_random' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a30b8e0> <<< 13118 1727204063.51344: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/hashlib.cpython-39.pyc matches /usr/lib64/python3.9/hashlib.py <<< 13118 1727204063.51348: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/hashlib.cpython-39.pyc' <<< 13118 1727204063.51617: stdout chunk (state=3): >>># extension module '_hashlib' loaded from '/usr/lib64/python3.9/lib-dynload/_hashlib.cpython-39-x86_64-linux-gnu.so' # extension module '_hashlib' executed from '/usr/lib64/python3.9/lib-dynload/_hashlib.cpython-39-x86_64-linux-gnu.so' import '_hashlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a30bd30> # extension module '_blake2' loaded from '/usr/lib64/python3.9/lib-dynload/_blake2.cpython-39-x86_64-linux-gnu.so' # extension module '_blake2' executed from '/usr/lib64/python3.9/lib-dynload/_blake2.cpython-39-x86_64-linux-gnu.so' import '_blake2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a315280> import 'hashlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a30b970> import 'random' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a2feac0> import 'tempfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a426640> <<< 13118 1727204063.51638: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/zipfile.cpython-39.pyc matches /usr/lib64/python3.9/zipfile.py <<< 13118 1727204063.51693: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/zipfile.cpython-39.pyc' <<< 13118 1727204063.51722: stdout chunk (state=3): >>>import 'zipfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a30bb20> <<< 13118 1727204063.51873: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/encodings/cp437.pyc' <<< 13118 1727204063.51885: stdout chunk (state=3): >>>import 'encodings.cp437' # <_frozen_importlib_external.SourcelessFileLoader object at 0x7f111a232700> <<< 13118 1727204063.52126: stdout chunk (state=3): >>># zipimport: found 103 names in '/tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip' <<< 13118 1727204063.52129: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.52213: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.52237: stdout chunk (state=3): >>>import ansible # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/__init__.py <<< 13118 1727204063.52255: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.52259: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.52292: stdout chunk (state=3): >>>import ansible.module_utils # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/__init__.py <<< 13118 1727204063.52295: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.53519: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.54436: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/__future__.cpython-39.pyc matches /usr/lib64/python3.9/__future__.py # code object from '/usr/lib64/python3.9/__pycache__/__future__.cpython-39.pyc' import '__future__' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a171850> <<< 13118 1727204063.54458: stdout chunk (state=3): >>># /usr/lib64/python3.9/json/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/json/__init__.py # code object from '/usr/lib64/python3.9/json/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204063.54476: stdout chunk (state=3): >>># /usr/lib64/python3.9/json/__pycache__/decoder.cpython-39.pyc matches /usr/lib64/python3.9/json/decoder.py # code object from '/usr/lib64/python3.9/json/__pycache__/decoder.cpython-39.pyc' <<< 13118 1727204063.54500: stdout chunk (state=3): >>># /usr/lib64/python3.9/json/__pycache__/scanner.cpython-39.pyc matches /usr/lib64/python3.9/json/scanner.py # code object from '/usr/lib64/python3.9/json/__pycache__/scanner.cpython-39.pyc' <<< 13118 1727204063.54531: stdout chunk (state=3): >>># extension module '_json' loaded from '/usr/lib64/python3.9/lib-dynload/_json.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204063.54536: stdout chunk (state=3): >>># extension module '_json' executed from '/usr/lib64/python3.9/lib-dynload/_json.cpython-39-x86_64-linux-gnu.so' import '_json' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a171160> <<< 13118 1727204063.54555: stdout chunk (state=3): >>>import 'json.scanner' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a171280> <<< 13118 1727204063.54588: stdout chunk (state=3): >>>import 'json.decoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a171fa0> <<< 13118 1727204063.54618: stdout chunk (state=3): >>># /usr/lib64/python3.9/json/__pycache__/encoder.cpython-39.pyc matches /usr/lib64/python3.9/json/encoder.py <<< 13118 1727204063.54621: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/json/__pycache__/encoder.cpython-39.pyc' <<< 13118 1727204063.54673: stdout chunk (state=3): >>>import 'json.encoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1714f0> <<< 13118 1727204063.54676: stdout chunk (state=3): >>>import 'json' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a171dc0> import 'atexit' # <<< 13118 1727204063.54692: stdout chunk (state=3): >>># extension module 'fcntl' loaded from '/usr/lib64/python3.9/lib-dynload/fcntl.cpython-39-x86_64-linux-gnu.so' # extension module 'fcntl' executed from '/usr/lib64/python3.9/lib-dynload/fcntl.cpython-39-x86_64-linux-gnu.so' import 'fcntl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a171580> <<< 13118 1727204063.54713: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/locale.cpython-39.pyc matches /usr/lib64/python3.9/locale.py <<< 13118 1727204063.54734: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/locale.cpython-39.pyc' <<< 13118 1727204063.54772: stdout chunk (state=3): >>>import 'locale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a171100> <<< 13118 1727204063.54800: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/platform.cpython-39.pyc matches /usr/lib64/python3.9/platform.py <<< 13118 1727204063.54803: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/platform.cpython-39.pyc' <<< 13118 1727204063.54821: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/subprocess.cpython-39.pyc matches /usr/lib64/python3.9/subprocess.py <<< 13118 1727204063.54846: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/subprocess.cpython-39.pyc' <<< 13118 1727204063.54866: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/signal.cpython-39.pyc matches /usr/lib64/python3.9/signal.py # code object from '/usr/lib64/python3.9/__pycache__/signal.cpython-39.pyc' <<< 13118 1727204063.54949: stdout chunk (state=3): >>>import 'signal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1470a0> <<< 13118 1727204063.54984: stdout chunk (state=3): >>># extension module '_posixsubprocess' loaded from '/usr/lib64/python3.9/lib-dynload/_posixsubprocess.cpython-39-x86_64-linux-gnu.so' # extension module '_posixsubprocess' executed from '/usr/lib64/python3.9/lib-dynload/_posixsubprocess.cpython-39-x86_64-linux-gnu.so' import '_posixsubprocess' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b08370> <<< 13118 1727204063.55013: stdout chunk (state=3): >>># extension module 'select' loaded from '/usr/lib64/python3.9/lib-dynload/select.cpython-39-x86_64-linux-gnu.so' # extension module 'select' executed from '/usr/lib64/python3.9/lib-dynload/select.cpython-39-x86_64-linux-gnu.so' import 'select' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b08070> <<< 13118 1727204063.55037: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/selectors.cpython-39.pyc matches /usr/lib64/python3.9/selectors.py <<< 13118 1727204063.55045: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/selectors.cpython-39.pyc' <<< 13118 1727204063.55084: stdout chunk (state=3): >>>import 'selectors' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119b08cd0> <<< 13118 1727204063.55087: stdout chunk (state=3): >>>import 'subprocess' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a159dc0> <<< 13118 1727204063.55251: stdout chunk (state=3): >>>import 'platform' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1593a0> <<< 13118 1727204063.55287: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/shlex.cpython-39.pyc matches /usr/lib64/python3.9/shlex.py <<< 13118 1727204063.55290: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/shlex.cpython-39.pyc' <<< 13118 1727204063.55294: stdout chunk (state=3): >>>import 'shlex' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a159f40> <<< 13118 1727204063.55312: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/traceback.cpython-39.pyc matches /usr/lib64/python3.9/traceback.py # code object from '/usr/lib64/python3.9/__pycache__/traceback.cpython-39.pyc' <<< 13118 1727204063.55351: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/linecache.cpython-39.pyc matches /usr/lib64/python3.9/linecache.py # code object from '/usr/lib64/python3.9/__pycache__/linecache.cpython-39.pyc' <<< 13118 1727204063.55380: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/tokenize.cpython-39.pyc matches /usr/lib64/python3.9/tokenize.py <<< 13118 1727204063.55383: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/tokenize.cpython-39.pyc' <<< 13118 1727204063.55410: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/token.cpython-39.pyc matches /usr/lib64/python3.9/token.py # code object from '/usr/lib64/python3.9/__pycache__/token.cpython-39.pyc' <<< 13118 1727204063.55413: stdout chunk (state=3): >>>import 'token' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1a6f40> <<< 13118 1727204063.55794: stdout chunk (state=3): >>>import 'tokenize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a173d60> import 'linecache' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a173430> import 'traceback' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119be1af0> # extension module 'syslog' loaded from '/usr/lib64/python3.9/lib-dynload/syslog.cpython-39-x86_64-linux-gnu.so' # extension module 'syslog' executed from '/usr/lib64/python3.9/lib-dynload/syslog.cpython-39-x86_64-linux-gnu.so' import 'syslog' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a173550> # /usr/lib64/python3.9/site-packages/systemd/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/__init__.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/__init__.cpython-39.pyc' import 'systemd' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a173580> # /usr/lib64/python3.9/site-packages/systemd/__pycache__/journal.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/journal.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/journal.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/datetime.cpython-39.pyc matches /usr/lib64/python3.9/datetime.py # code object from '/usr/lib64/python3.9/__pycache__/datetime.cpython-39.pyc' # extension module '_datetime' loaded from '/usr/lib64/python3.9/lib-dynload/_datetime.cpython-39-x86_64-linux-gnu.so' # extension module '_datetime' executed from '/usr/lib64/python3.9/lib-dynload/_datetime.cpython-39-x86_64-linux-gnu.so' import '_datetime' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b76fa0> import 'datetime' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1b8280> # /usr/lib64/python3.9/__pycache__/uuid.cpython-39.pyc matches /usr/lib64/python3.9/uuid.py # code object from '/usr/lib64/python3.9/__pycache__/uuid.cpython-39.pyc' # extension module '_uuid' loaded from '/usr/lib64/python3.9/lib-dynload/_uuid.cpython-39-x86_64-linux-gnu.so' # extension module '_uuid' executed from '/usr/lib64/python3.9/lib-dynload/_uuid.cpython-39-x86_64-linux-gnu.so' import '_uuid' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b73820> <<< 13118 1727204063.55798: stdout chunk (state=3): >>>import 'uuid' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1b8400> <<< 13118 1727204063.55811: stdout chunk (state=3): >>># /usr/lib64/python3.9/logging/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/logging/__init__.py <<< 13118 1727204063.55895: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/logging/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/string.cpython-39.pyc matches /usr/lib64/python3.9/string.py # code object from '/usr/lib64/python3.9/__pycache__/string.cpython-39.pyc' import '_string' # <<< 13118 1727204063.55934: stdout chunk (state=3): >>>import 'string' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1b8c40> <<< 13118 1727204063.56057: stdout chunk (state=3): >>>import 'logging' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119b737c0> <<< 13118 1727204063.56152: stdout chunk (state=3): >>># extension module 'systemd._journal' loaded from '/usr/lib64/python3.9/site-packages/systemd/_journal.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._journal' executed from '/usr/lib64/python3.9/site-packages/systemd/_journal.cpython-39-x86_64-linux-gnu.so' import 'systemd._journal' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a1511c0> <<< 13118 1727204063.56190: stdout chunk (state=3): >>># extension module 'systemd._reader' loaded from '/usr/lib64/python3.9/site-packages/systemd/_reader.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204063.56193: stdout chunk (state=3): >>># extension module 'systemd._reader' executed from '/usr/lib64/python3.9/site-packages/systemd/_reader.cpython-39-x86_64-linux-gnu.so' import 'systemd._reader' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a1b89d0> <<< 13118 1727204063.56225: stdout chunk (state=3): >>># extension module 'systemd.id128' loaded from '/usr/lib64/python3.9/site-packages/systemd/id128.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd.id128' executed from '/usr/lib64/python3.9/site-packages/systemd/id128.cpython-39-x86_64-linux-gnu.so' import 'systemd.id128' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a1b8550> import 'systemd.journal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1b1940> <<< 13118 1727204063.56254: stdout chunk (state=3): >>># /usr/lib64/python3.9/site-packages/systemd/__pycache__/daemon.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/daemon.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/daemon.cpython-39.pyc' <<< 13118 1727204063.56284: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/socket.cpython-39.pyc matches /usr/lib64/python3.9/socket.py <<< 13118 1727204063.56287: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/socket.cpython-39.pyc' <<< 13118 1727204063.56331: stdout chunk (state=3): >>># extension module '_socket' loaded from '/usr/lib64/python3.9/lib-dynload/_socket.cpython-39-x86_64-linux-gnu.so' # extension module '_socket' executed from '/usr/lib64/python3.9/lib-dynload/_socket.cpython-39-x86_64-linux-gnu.so' import '_socket' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b68910> <<< 13118 1727204063.56503: stdout chunk (state=3): >>># extension module 'array' loaded from '/usr/lib64/python3.9/lib-dynload/array.cpython-39-x86_64-linux-gnu.so' # extension module 'array' executed from '/usr/lib64/python3.9/lib-dynload/array.cpython-39-x86_64-linux-gnu.so' import 'array' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b85dc0> <<< 13118 1727204063.56523: stdout chunk (state=3): >>>import 'socket' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119b72550> <<< 13118 1727204063.56557: stdout chunk (state=3): >>># extension module 'systemd._daemon' loaded from '/usr/lib64/python3.9/site-packages/systemd/_daemon.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._daemon' executed from '/usr/lib64/python3.9/site-packages/systemd/_daemon.cpython-39-x86_64-linux-gnu.so' import 'systemd._daemon' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b68eb0> import 'systemd.daemon' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119b72970> <<< 13118 1727204063.56562: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.56594: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.56598: stdout chunk (state=3): >>>import ansible.module_utils.compat # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/compat/__init__.py <<< 13118 1727204063.56600: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.56673: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.56754: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.56757: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.56784: stdout chunk (state=3): >>>import ansible.module_utils.common # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/__init__.py # zipimport: zlib available <<< 13118 1727204063.56789: stdout chunk (state=3): >>># zipimport: zlib available import ansible.module_utils.common.text # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/text/__init__.py <<< 13118 1727204063.56802: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.56891: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.56991: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.57434: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.57904: stdout chunk (state=3): >>>import ansible.module_utils.six # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/six/__init__.py import 'ansible.module_utils.six.moves' # <<< 13118 1727204063.57908: stdout chunk (state=3): >>>import 'ansible.module_utils.six.moves.collections_abc' # import ansible.module_utils.common.text.converters # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/text/converters.py <<< 13118 1727204063.57930: stdout chunk (state=3): >>># /usr/lib64/python3.9/ctypes/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/ctypes/__init__.py <<< 13118 1727204063.57933: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/ctypes/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204063.57992: stdout chunk (state=3): >>># extension module '_ctypes' loaded from '/usr/lib64/python3.9/lib-dynload/_ctypes.cpython-39-x86_64-linux-gnu.so' # extension module '_ctypes' executed from '/usr/lib64/python3.9/lib-dynload/_ctypes.cpython-39-x86_64-linux-gnu.so' import '_ctypes' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119baf7f0> <<< 13118 1727204063.58068: stdout chunk (state=3): >>># /usr/lib64/python3.9/ctypes/__pycache__/_endian.cpython-39.pyc matches /usr/lib64/python3.9/ctypes/_endian.py # code object from '/usr/lib64/python3.9/ctypes/__pycache__/_endian.cpython-39.pyc' <<< 13118 1727204063.58071: stdout chunk (state=3): >>>import 'ctypes._endian' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119bb48b0> <<< 13118 1727204063.58088: stdout chunk (state=3): >>>import 'ctypes' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111971a940> <<< 13118 1727204063.58124: stdout chunk (state=3): >>>import ansible.module_utils.compat.selinux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/compat/selinux.py <<< 13118 1727204063.58127: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.58160: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.58169: stdout chunk (state=3): >>>import ansible.module_utils._text # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/_text.py <<< 13118 1727204063.58173: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.58292: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.58445: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/copy.cpython-39.pyc matches /usr/lib64/python3.9/copy.py # code object from '/usr/lib64/python3.9/__pycache__/copy.cpython-39.pyc' <<< 13118 1727204063.58448: stdout chunk (state=3): >>>import 'copy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119bec730> <<< 13118 1727204063.58451: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.58883: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.59195: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.59250: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.59320: stdout chunk (state=3): >>>import ansible.module_utils.common.collections # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/collections.py # zipimport: zlib available <<< 13118 1727204063.59352: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.59394: stdout chunk (state=3): >>>import ansible.module_utils.common.warnings # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/warnings.py <<< 13118 1727204063.59397: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.59447: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.59552: stdout chunk (state=3): >>>import ansible.module_utils.errors # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/errors.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.parsing # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/parsing/__init__.py # zipimport: zlib available <<< 13118 1727204063.59592: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.59634: stdout chunk (state=3): >>>import ansible.module_utils.parsing.convert_bool # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/parsing/convert_bool.py <<< 13118 1727204063.59637: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.59814: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.60008: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/ast.cpython-39.pyc matches /usr/lib64/python3.9/ast.py <<< 13118 1727204063.60028: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/ast.cpython-39.pyc' <<< 13118 1727204063.60041: stdout chunk (state=3): >>>import '_ast' # <<< 13118 1727204063.60113: stdout chunk (state=3): >>>import 'ast' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1742e0> <<< 13118 1727204063.60120: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.60175: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.60248: stdout chunk (state=3): >>>import ansible.module_utils.common.text.formatters # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/text/formatters.py <<< 13118 1727204063.60253: stdout chunk (state=3): >>>import ansible.module_utils.common.validation # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/validation.py import ansible.module_utils.common.parameters # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/parameters.py import ansible.module_utils.common.arg_spec # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/arg_spec.py <<< 13118 1727204063.60274: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.60306: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.60343: stdout chunk (state=3): >>>import ansible.module_utils.common.locale # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/locale.py <<< 13118 1727204063.60347: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.60385: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.60488: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.60604: stdout chunk (state=3): >>># zipimport: zlib available # /usr/lib64/python3.9/site-packages/selinux/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/selinux/__init__.py # code object from '/usr/lib64/python3.9/site-packages/selinux/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204063.60683: stdout chunk (state=3): >>># extension module 'selinux._selinux' loaded from '/usr/lib64/python3.9/site-packages/selinux/_selinux.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204063.60686: stdout chunk (state=3): >>># extension module 'selinux._selinux' executed from '/usr/lib64/python3.9/site-packages/selinux/_selinux.cpython-39-x86_64-linux-gnu.so' import 'selinux._selinux' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119ba6880> <<< 13118 1727204063.60776: stdout chunk (state=3): >>>import 'selinux' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119596550> <<< 13118 1727204063.60813: stdout chunk (state=3): >>>import ansible.module_utils.common.file # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/file.py import ansible.module_utils.common.process # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/process.py <<< 13118 1727204063.60816: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.60870: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.60921: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.60945: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.60987: stdout chunk (state=3): >>># /usr/lib/python3.9/site-packages/__pycache__/distro.cpython-39.pyc matches /usr/lib/python3.9/site-packages/distro.py <<< 13118 1727204063.60999: stdout chunk (state=3): >>># code object from '/usr/lib/python3.9/site-packages/__pycache__/distro.cpython-39.pyc' <<< 13118 1727204063.61017: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/argparse.cpython-39.pyc matches /usr/lib64/python3.9/argparse.py <<< 13118 1727204063.61054: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/argparse.cpython-39.pyc' <<< 13118 1727204063.61074: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/gettext.cpython-39.pyc matches /usr/lib64/python3.9/gettext.py <<< 13118 1727204063.61091: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/gettext.cpython-39.pyc' <<< 13118 1727204063.61169: stdout chunk (state=3): >>>import 'gettext' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119bb7910> <<< 13118 1727204063.61209: stdout chunk (state=3): >>>import 'argparse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a143970> <<< 13118 1727204063.61274: stdout chunk (state=3): >>>import 'distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119bea850> <<< 13118 1727204063.61280: stdout chunk (state=3): >>># destroy ansible.module_utils.distro import ansible.module_utils.distro # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/distro/__init__.py # zipimport: zlib available <<< 13118 1727204063.61298: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.61329: stdout chunk (state=3): >>>import ansible.module_utils.common._utils # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/_utils.py <<< 13118 1727204063.61333: stdout chunk (state=3): >>>import ansible.module_utils.common.sys_info # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/sys_info.py <<< 13118 1727204063.61411: stdout chunk (state=3): >>>import ansible.module_utils.basic # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/basic.py <<< 13118 1727204063.61417: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.61420: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.61434: stdout chunk (state=3): >>>import ansible.modules # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/modules/__init__.py # zipimport: zlib available <<< 13118 1727204063.61561: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available # zipimport: zlib available <<< 13118 1727204063.61579: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.61616: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.61648: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.61683: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.61710: stdout chunk (state=3): >>>import ansible.module_utils.facts.namespace # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/namespace.py <<< 13118 1727204063.61723: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.61787: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.61850: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.61877: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.61899: stdout chunk (state=3): >>>import ansible.module_utils.compat.typing # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/compat/typing.py # zipimport: zlib available <<< 13118 1727204063.62049: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.62188: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.62225: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.62278: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/__init__.py <<< 13118 1727204063.62283: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204063.62304: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/context.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/context.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/context.cpython-39.pyc' <<< 13118 1727204063.62333: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/process.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/process.py <<< 13118 1727204063.62336: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/process.cpython-39.pyc' <<< 13118 1727204063.62352: stdout chunk (state=3): >>>import 'multiprocessing.process' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111949bc70> <<< 13118 1727204063.62380: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/reduction.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/reduction.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/reduction.cpython-39.pyc' <<< 13118 1727204063.62404: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/pickle.cpython-39.pyc matches /usr/lib64/python3.9/pickle.py <<< 13118 1727204063.62433: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/pickle.cpython-39.pyc' <<< 13118 1727204063.62463: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_compat_pickle.cpython-39.pyc matches /usr/lib64/python3.9/_compat_pickle.py <<< 13118 1727204063.62474: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/_compat_pickle.cpython-39.pyc' <<< 13118 1727204063.62476: stdout chunk (state=3): >>>import '_compat_pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11196fda30> <<< 13118 1727204063.62516: stdout chunk (state=3): >>># extension module '_pickle' loaded from '/usr/lib64/python3.9/lib-dynload/_pickle.cpython-39-x86_64-linux-gnu.so' # extension module '_pickle' executed from '/usr/lib64/python3.9/lib-dynload/_pickle.cpython-39-x86_64-linux-gnu.so' import '_pickle' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f11196fd9a0> <<< 13118 1727204063.62657: stdout chunk (state=3): >>>import 'pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119743b20> <<< 13118 1727204063.62661: stdout chunk (state=3): >>>import 'multiprocessing.reduction' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119743550> <<< 13118 1727204063.62663: stdout chunk (state=3): >>>import 'multiprocessing.context' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111972f2e0> <<< 13118 1727204063.62679: stdout chunk (state=3): >>>import 'multiprocessing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111972f970> <<< 13118 1727204063.62682: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/pool.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/pool.py <<< 13118 1727204063.62684: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/pool.cpython-39.pyc' <<< 13118 1727204063.62687: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/queue.cpython-39.pyc matches /usr/lib64/python3.9/queue.py # code object from '/usr/lib64/python3.9/__pycache__/queue.cpython-39.pyc' <<< 13118 1727204063.62723: stdout chunk (state=3): >>># extension module '_queue' loaded from '/usr/lib64/python3.9/lib-dynload/_queue.cpython-39-x86_64-linux-gnu.so' # extension module '_queue' executed from '/usr/lib64/python3.9/lib-dynload/_queue.cpython-39-x86_64-linux-gnu.so' import '_queue' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f11196e02b0> import 'queue' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11196e0a00> <<< 13118 1727204063.62758: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/util.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/util.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/util.cpython-39.pyc' <<< 13118 1727204063.62794: stdout chunk (state=3): >>>import 'multiprocessing.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11196e0940> <<< 13118 1727204063.62844: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/connection.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/connection.py <<< 13118 1727204063.62849: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/connection.cpython-39.pyc' <<< 13118 1727204063.62871: stdout chunk (state=3): >>># extension module '_multiprocessing' loaded from '/usr/lib64/python3.9/lib-dynload/_multiprocessing.cpython-39-x86_64-linux-gnu.so' # extension module '_multiprocessing' executed from '/usr/lib64/python3.9/lib-dynload/_multiprocessing.cpython-39-x86_64-linux-gnu.so' import '_multiprocessing' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f11194fc0d0> <<< 13118 1727204063.62916: stdout chunk (state=3): >>>import 'multiprocessing.connection' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119ba23a0> <<< 13118 1727204063.62959: stdout chunk (state=3): >>>import 'multiprocessing.pool' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111972f670> import ansible.module_utils.facts.timeout # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/timeout.py <<< 13118 1727204063.62979: stdout chunk (state=3): >>>import ansible.module_utils.facts.collector # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/collector.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.other # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/other/__init__.py <<< 13118 1727204063.62982: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.63019: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.63078: stdout chunk (state=3): >>>import ansible.module_utils.facts.other.facter # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/other/facter.py <<< 13118 1727204063.63081: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.63111: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.63197: stdout chunk (state=3): >>>import ansible.module_utils.facts.other.ohai # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/other/ohai.py # zipimport: zlib available <<< 13118 1727204063.63220: stdout chunk (state=3): >>># zipimport: zlib available import ansible.module_utils.facts.system # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/__init__.py # zipimport: zlib available # zipimport: zlib available <<< 13118 1727204063.63251: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.apparmor # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/apparmor.py # zipimport: zlib available <<< 13118 1727204063.63297: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.63344: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.caps # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/caps.py <<< 13118 1727204063.63356: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.63379: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.63421: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.chroot # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/chroot.py # zipimport: zlib available <<< 13118 1727204063.63484: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.63529: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.63582: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.63634: stdout chunk (state=3): >>>import ansible.module_utils.facts.utils # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/utils.py import ansible.module_utils.facts.system.cmdline # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/cmdline.py # zipimport: zlib available <<< 13118 1727204063.64030: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.64407: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.distribution # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/distribution.py # zipimport: zlib available <<< 13118 1727204063.64448: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.64506: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.64537: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.64586: stdout chunk (state=3): >>>import ansible.module_utils.compat.datetime # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/compat/datetime.py import ansible.module_utils.facts.system.date_time # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/date_time.py <<< 13118 1727204063.64589: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.64636: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.64651: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.env # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/env.py # zipimport: zlib available <<< 13118 1727204063.64702: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.64761: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.dns # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/dns.py <<< 13118 1727204063.64773: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.64786: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.64819: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.fips # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/fips.py # zipimport: zlib available <<< 13118 1727204063.64850: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.64894: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.loadavg # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/loadavg.py <<< 13118 1727204063.64898: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.64952: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.65034: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/glob.cpython-39.pyc matches /usr/lib64/python3.9/glob.py # code object from '/usr/lib64/python3.9/__pycache__/glob.cpython-39.pyc' <<< 13118 1727204063.65057: stdout chunk (state=3): >>>import 'glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11193eaeb0> <<< 13118 1727204063.65085: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/configparser.cpython-39.pyc matches /usr/lib64/python3.9/configparser.py <<< 13118 1727204063.65109: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/configparser.cpython-39.pyc' <<< 13118 1727204063.65280: stdout chunk (state=3): >>>import 'configparser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11193ea9d0> import ansible.module_utils.facts.system.local # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/local.py <<< 13118 1727204063.65284: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.65338: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.65404: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.lsb # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/lsb.py <<< 13118 1727204063.65407: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.65479: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.65563: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.pkg_mgr # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/pkg_mgr.py # zipimport: zlib available <<< 13118 1727204063.65623: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.65693: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.platform # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/platform.py <<< 13118 1727204063.65706: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.65736: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.65785: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/ssl.cpython-39.pyc matches /usr/lib64/python3.9/ssl.py <<< 13118 1727204063.65806: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/ssl.cpython-39.pyc' <<< 13118 1727204063.65954: stdout chunk (state=3): >>># extension module '_ssl' loaded from '/usr/lib64/python3.9/lib-dynload/_ssl.cpython-39-x86_64-linux-gnu.so' # extension module '_ssl' executed from '/usr/lib64/python3.9/lib-dynload/_ssl.cpython-39-x86_64-linux-gnu.so' import '_ssl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119458bb0> <<< 13118 1727204063.66209: stdout chunk (state=3): >>>import 'ssl' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11193fba60> import ansible.module_utils.facts.system.python # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/python.py <<< 13118 1727204063.66213: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.66258: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.66317: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.selinux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/selinux.py <<< 13118 1727204063.66320: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.66386: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.66458: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.66552: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.66694: stdout chunk (state=3): >>>import ansible.module_utils.compat.version # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/compat/version.py import ansible.module_utils.facts.system.service_mgr # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/service_mgr.py # zipimport: zlib available <<< 13118 1727204063.66772: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.66786: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.ssh_pub_keys # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/ssh_pub_keys.py # zipimport: zlib available <<< 13118 1727204063.66819: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.66870: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/getpass.cpython-39.pyc matches /usr/lib64/python3.9/getpass.py # code object from '/usr/lib64/python3.9/__pycache__/getpass.cpython-39.pyc' <<< 13118 1727204063.66935: stdout chunk (state=3): >>># extension module 'termios' loaded from '/usr/lib64/python3.9/lib-dynload/termios.cpython-39-x86_64-linux-gnu.so' # extension module 'termios' executed from '/usr/lib64/python3.9/lib-dynload/termios.cpython-39-x86_64-linux-gnu.so' import 'termios' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111945e040> import 'getpass' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111945e6d0> <<< 13118 1727204063.66951: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.user # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/user.py <<< 13118 1727204063.66983: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/__init__.py <<< 13118 1727204063.66986: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.67019: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.67063: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.base # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/base.py # zipimport: zlib available <<< 13118 1727204063.67195: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.67324: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.aix # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/aix.py # zipimport: zlib available <<< 13118 1727204063.67416: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.67494: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.67536: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.67589: stdout chunk (state=3): >>>import ansible.module_utils.facts.sysctl # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/sysctl.py import ansible.module_utils.facts.hardware.darwin # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/darwin.py <<< 13118 1727204063.67593: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.67688: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.67701: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.67816: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.67945: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.freebsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/freebsd.py import ansible.module_utils.facts.hardware.dragonfly # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/dragonfly.py # zipimport: zlib available <<< 13118 1727204063.68053: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.68169: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.hpux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/hpux.py <<< 13118 1727204063.68172: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.68196: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.68232: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.68676: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.69113: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.linux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/linux.py import ansible.module_utils.facts.hardware.hurd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/hurd.py <<< 13118 1727204063.69117: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.69198: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.69290: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.netbsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/netbsd.py # zipimport: zlib available <<< 13118 1727204063.69383: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.69463: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.openbsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/openbsd.py <<< 13118 1727204063.69483: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.69601: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.69772: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.sunos # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/sunos.py # zipimport: zlib available <<< 13118 1727204063.69775: stdout chunk (state=3): >>># zipimport: zlib available import ansible.module_utils.facts.network # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/__init__.py <<< 13118 1727204063.69788: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.69814: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.69859: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.base # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/base.py # zipimport: zlib available <<< 13118 1727204063.69948: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.70032: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.70206: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.70392: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.generic_bsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/generic_bsd.py import ansible.module_utils.facts.network.aix # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/aix.py <<< 13118 1727204063.70395: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.70426: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.70462: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.darwin # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/darwin.py # zipimport: zlib available <<< 13118 1727204063.70497: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.70527: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.dragonfly # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/dragonfly.py <<< 13118 1727204063.70532: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.70585: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.70651: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.fc_wwn # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/fc_wwn.py # zipimport: zlib available <<< 13118 1727204063.70678: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.70712: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.freebsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/freebsd.py <<< 13118 1727204063.70715: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.70761: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.70816: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.hpux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/hpux.py # zipimport: zlib available <<< 13118 1727204063.70879: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.70938: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.hurd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/hurd.py <<< 13118 1727204063.70942: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.71147: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.71376: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.linux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/linux.py # zipimport: zlib available <<< 13118 1727204063.71433: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.71490: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.iscsi # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/iscsi.py # zipimport: zlib available <<< 13118 1727204063.71526: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.71559: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.nvme # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/nvme.py # zipimport: zlib available <<< 13118 1727204063.71596: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.71653: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.netbsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/netbsd.py <<< 13118 1727204063.71657: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.71705: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.71708: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.openbsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/openbsd.py # zipimport: zlib available <<< 13118 1727204063.71776: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.71878: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.sunos # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/sunos.py # zipimport: zlib available # zipimport: zlib available <<< 13118 1727204063.71894: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/__init__.py # zipimport: zlib available <<< 13118 1727204063.71935: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.71987: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.base # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/base.py <<< 13118 1727204063.72010: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13118 1727204063.72024: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72063: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72104: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72161: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72242: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.sysctl # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/sysctl.py import ansible.module_utils.facts.virtual.freebsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/freebsd.py import ansible.module_utils.facts.virtual.dragonfly # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/dragonfly.py <<< 13118 1727204063.72256: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72296: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72344: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.hpux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/hpux.py <<< 13118 1727204063.72357: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72513: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72677: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.linux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/linux.py <<< 13118 1727204063.72689: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72725: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72763: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.netbsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/netbsd.py <<< 13118 1727204063.72778: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72804: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72853: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.openbsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/openbsd.py # zipimport: zlib available <<< 13118 1727204063.72922: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.72998: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.sunos # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/sunos.py import ansible.module_utils.facts.default_collectors # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/default_collectors.py # zipimport: zlib available <<< 13118 1727204063.73078: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.73153: stdout chunk (state=3): >>>import ansible.module_utils.facts.ansible_collector # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/ansible_collector.py import ansible.module_utils.facts.compat # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/compat.py import ansible.module_utils.facts # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/__init__.py <<< 13118 1727204063.73232: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204063.73425: stdout chunk (state=3): >>>import 'gc' # <<< 13118 1727204063.73770: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/idna.cpython-39.pyc matches /usr/lib64/python3.9/encodings/idna.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/idna.cpython-39.pyc' <<< 13118 1727204063.73803: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/stringprep.cpython-39.pyc matches /usr/lib64/python3.9/stringprep.py # code object from '/usr/lib64/python3.9/__pycache__/stringprep.cpython-39.pyc' <<< 13118 1727204063.73856: stdout chunk (state=3): >>># extension module 'unicodedata' loaded from '/usr/lib64/python3.9/lib-dynload/unicodedata.cpython-39-x86_64-linux-gnu.so' # extension module 'unicodedata' executed from '/usr/lib64/python3.9/lib-dynload/unicodedata.cpython-39-x86_64-linux-gnu.so' import 'unicodedata' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f11193e0190> import 'stringprep' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11193e01c0> <<< 13118 1727204063.73920: stdout chunk (state=3): >>>import 'encodings.idna' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111938ce80> <<< 13118 1727204063.75843: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/queues.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/queues.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/queues.cpython-39.pyc' <<< 13118 1727204063.75884: stdout chunk (state=3): >>>import 'multiprocessing.queues' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11193e02b0> <<< 13118 1727204063.75888: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/synchronize.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/synchronize.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/synchronize.cpython-39.pyc' <<< 13118 1727204063.75903: stdout chunk (state=3): >>>import 'multiprocessing.synchronize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11193a2730> <<< 13118 1727204063.75946: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/dummy/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/dummy/__init__.py # code object from '/usr/lib64/python3.9/multiprocessing/dummy/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204063.75997: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/dummy/__pycache__/connection.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/dummy/connection.py # code object from '/usr/lib64/python3.9/multiprocessing/dummy/__pycache__/connection.cpython-39.pyc' import 'multiprocessing.dummy.connection' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11191f2280> <<< 13118 1727204063.76013: stdout chunk (state=3): >>>import 'multiprocessing.dummy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11191f2070> <<< 13118 1727204063.76273: stdout chunk (state=3): >>>PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame <<< 13118 1727204064.01074: stdout chunk (state=3): >>> {"ansible_facts": {"ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "root", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_system": "Linux", "ansible_kernel": "5.14.0-511.el9.x86_64", "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Thu Sep 19 06:52:39 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.9.19", "ansible_fqdn": "managed-node2", "ansible_hostname": "managed-node2", "ansible_nodename": "managed-node2", "ansible_domain": "", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "e28ab0e542474a869c23f7ace4640799", "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_python": {"version": {"major": 3, "minor": 9, "micro": 19, "releaselevel": "final", "serial": 0}, "version_info": [3, 9, 19, "final", 0], "executable": "/usr/bin/python3.9", "has_sslcontext": true, "type": "cpython"}, "ansible_fibre_channel_wwn": [], "ansible_env": {"PYTHONVERBOSE": "1", "SHELL": "/bin/bash", "PWD": "/root", "LOGNAME": "root", "XDG_SESSION_TYPE": "tty", "_": "/usr/bin/python3.9", "MOTD_SHOWN": "pam", "HOME": "/root", "LANG": "en_US.UTF-8", "LS_COLORS": "", "SSH_CONNECTION": "10.31.14.85 48676 10.31.13.78 22", "XDG_SESSION_CLASS": "user", "SELINUX_ROLE_REQUESTED": "", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "USER": "root", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "1", "XDG_SESSION_ID": "5", "XDG_RUNTIME_DIR": "/run/user/0", "SSH_CLIENT": "10.31.14.85 48676 22", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "which_declare": "declare -f", "PATH": "/root/.local/bin:/root/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "SELINUX_LEVEL_REQUESTED": "", "DBUS_SESSION_BUS_ADDR<<< 13118 1727204064.01144: stdout chunk (state=3): >>>ESS": "unix:path=/run/user/0/bus", "SSH_TTY": "/dev/pts/0", "BASH_FUNC_which%%": "() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz", "1", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz"], "ansible_processor_count": 1, "ansible_processor_cores": 1, "ansible_processor_threads_per_core": 2, "ansible_processor_vcpus": 2, "ansible_processor_nproc": 2, "ansible_memtotal_mb": 3532, "ansible_memfree_mb": 2785, "ansible_swaptotal_mb": 0, "ansible_swapfree_mb": 0, "ansible_memory_mb": {"real": {"total": 3532, "used": 747, "free": 2785}, "nocache": {"free": 3253, "used": 279}, "swap": {"total": 0, "free": 0, "used": 0, "cached": 0}}, "ansible_bios_date": "08/24/2006", "ansible_bios_vendor": "Xen", "ansible_bios_version": "4.11.amazon", "ansible_board_asset_tag": "NA", "ansible_board_name": "NA", "ansible_board_serial": "NA", "ansible_board_vendor": "NA", "ansible_board_version": "NA", "ansible_chassis_asset_tag": "NA", "ansible_chassis_serial": "NA", "ansible_chassis_vendor": "Xen", "ansible_chassis_version": "NA", "ansible_form_factor": "Other", "ansible_product_name": "HVM domU", "ansible_product_serial": "ec243623-fa66-7445-44ba-1070930583a9", "ansible_product_uuid": "ec243623-fa66-7445-44ba-1070930583a9", "ansible_product_version": "4.11.amazon", "ansible_system_vendor": "Xen", "ansible_devices": {"xvda": {"virtual": 1, "links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "vendor": null, "model": null, "sas_address": null, "sas_device_handle": null, "removable": "0", "support_discard": "512", "partitions": {"xvda1": {"links": {"ids": [], "uuids": ["ad406aa3-aab4-4a6a-aa73-3e870a6316ae"], "labels": [], "masters": []}, "start": "2048", "sectors": "524285919", "sectorsize": 512, "size": "250.00 GB", "uuid": "ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "holders": []}}, "rotational": "0", "scheduler_mode": "mq-deadline", "sectors": "524288000", "sectorsize": "512", "size": "250.00 GB", "host": "", "holders": []}}, "ansible_device_links": {"ids": {}, "uuids": {"xvda1": ["ad406aa3-aab4-4a6a-aa73-3e870a6316ae"]}, "labels": {}, "masters": {}}, "ansible_uptime_seconds": 426, "ansible_lvm": "N/A", "ansible_mounts": [{"mount": "/", "device": "/dev/xvda1", "fstype": "xfs", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "dump": 0, "passno": 0, "size_total": 268367278080, "size_available": 264273084416, "block_size": 4096, "block_total": 65519355, "block_available": 64519796, "block_used": 999559, "inode_total": 131071472, "inode_available": 130998253, "inode_used": 73219, "uuid": "ad406aa3-aab4-4a6a-aa73-3e870a6316ae"}], "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAPleAC0mV69PNpLSbmzZvoLD9LsCBzX6IHRLXV1uktk0r66T6Y57EoVgflJTdo6yU0zTaJjonNzFmvC69tiRsCyywGjnvnBOvIH2vrgNGCUdVYPZbbtmQlJvol7NFFfyXQR4RSPqBKT67rYbCzbETM4j+bdDgTeDk6l7wXwz9RVvAAAAFQCuAyyjbOBDKyIW26LGcI9/nmWpHwAAAIEApIE1W6KQ7qs5kJXBdSaPoWaZUxuQhXkPWORFe7/MBn5SojDfxvJjFPo6t4QsovaCnm532Zghh1ZdB0pNm0vYcRbz3wMdfMucw/KHWt6ZEtI+sLwuMyhAVEXzmE34iXkyePtELiYzY6NyxuJ04IujI9UwD7ZnqFBHVFz529oXikIAAACBAPdUu+4Qo82CMcmrGD9vNUgtsts6GCjqBDuov8GJEALZ9ZNLlyVoNtBHLMQH9e0czLygyNGw/IDosRQkKdX4Vh4A7KXujTIOyytaN4JVJCuOBY/PeX4lreAO/UTTUJ27yT/J0Oy2Hbt+d8fZnTkZReRNPFCzvdb1nuPMG5nAyQtL", "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQCzkKXWiNuOrU77QQcZuT2T9XVh655Sh8Sv9vLWLa1uj7ceaNsB0TBiqvDFvYPENhdKceYaGAFU7sjqbmp5dlivYwPBiBWvcOgqnpBqrMG5SvP1RMiORpW6GupBLnUaMVjopPLIi0/CDlSl2eODcEnQI6BpxCCSedEKU9UrRrCFJy+6KPQXepPwKwPTd1TMzO8wpo57B5MYrjnquTNxMfgBkYsHB/V77d0tKq8qGBTkAPD8wEWLIcZOI+SyYEfCraQ95dOGAPRTFijnd7S15CugSlJ/vvcHSFXOlbgFzeNnU2jZneagkBfaOJch72opD3ebISSHCx1/kJvHN7MbksI+ljJa3Nw5LwP1XjUpT7dQMOZJDdVStXKp86K4XpWud+wMbQVVyU5QoFsCl7YTWWmSDRiPJOQI2myfizCT8i42rJ0WXm5OnqpHn1Jw4nGlcVnfgPQA/zxMldzReXdHnvriqKC9+97XgY6pj42YYP78PhOu1D2xH1AXmloNM+63VvU=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPT1h7wNcUomxtav688iXvnCnFqrHnEKf4gRaBY3w4BwbWOGxE8hq5snF9Tp+0agFeN/u980/y8BJWdWIO9Lz8I=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIPe8liWy3mh5GzCz9W616J2ArXnLOjLOZSwfmBX3Q1SI", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "9", "ansible_distribution_major_version": "9", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_is_chroot": false, "ansible_apparmor": {"status": "disabled"}, "ansible_iscsi_iqn": "", "ansible_virtualization_type": "xen", "ansible_virtualization_role": "guest", "ansible_virtualization_tech_guest": ["xen"], "ansible_virtualization_tech_host": [], "ansible_hostnqn": "nqn.2014-08.org.nvmexpress:uuid:d5aef1ea-3141-48ae-bf33-0c6b351dd422", "ansible_fips": false, "ansible_pkg_mgr": "dnf", "ansible_service_mgr": "systemd", "ansible_loadavg": {"1m": 0.38, "5m": 0.26, "15m": 0.13}, "ansible_interfaces": ["lo", "eth0"], "ansible_eth0": {"device": "eth0", "macaddress": "0a:ff:ff:f5:f2:b9", "mtu": 9001, "active": true, "module": "xen_netfront", "type": "ether", "pciid": "vif-0", "promisc": false, "ipv4": {"address": "10.31.13.78", "broadcast": "10.31.15.255", "netmask": "255.255.252.0", "network": "10.31.12.0", "prefix": "22"}, "ipv6": [{"address": "fe80::8ff:ffff:fef5:f2b9", "prefix": "64", "scope": "link"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "on [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv6": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on", "tx_scatter_gather_fraglist": "off [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "off [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "off [fixed]", "tx_lockless": "off [fixed]", "netns_local": "off [fixed]", "tx_gso_robust": "on [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_gso_list": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off", "loopback": "off [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_lo": {"device": "lo", "mtu": 65536, "active": true, "type": "loopback", "promisc": false, "ipv4": {"address": "<<< 13118 1727204064.01154: stdout chunk (state=3): >>>127.0.0.1", "broadcast": "", "netmask": "255.0.0.0", "network": "127.0.0.0", "prefix": "8"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "on [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "on [fixed]", "tx_lockless": "on [fixed]", "netns_local": "on [fixed]", "tx_gso_robust": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "on", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "on", "tx_gso_list": "on", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off [fixed]", "loopback": "on [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_default_ipv4": {"gateway": "10.31.12.1", "interface": "eth0", "address": "10.31.13.78", "broadcast": "10.31.15.255", "netmask": "255.255.252.0", "network": "10.31.12.0", "prefix": "22", "macaddress": "0a:ff:ff:f5:f2:b9", "mtu": 9001, "type": "ether", "alias": "eth0"}, "ansible_default_ipv6": {}, "ansible_all_ipv4_addresses": ["10.31.13.78"], "ansible_all_ipv6_addresses": ["fe80::8ff:ffff:fef5:f2b9"], "ansible_locally_reachable_ips": {"ipv4": ["10.31.13.78", "127.0.0.0/8", "127.0.0.1"], "ipv6": ["::1", "fe80::8ff:ffff:fef5:f2b9"]}, "ansible_lsb": {}, "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-5.14.0-511.el9.x86_64", "root": "UUID=ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": "ttyS0,115200n8"}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-5.14.0-511.el9.x86_64", "root": "UUID=ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": ["tty0", "ttyS0,115200n8"]}, "ansible_local": {}, "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Tuesday", "weekday_number": "2", "weeknumber": "39", "day": "24", "hour": "14", "minute": "54", "second": "24", "epoch": "1727204064", "epoch_int": "1727204064", "date": "2024-09-24", "time": "14:54:24", "iso8601_micro": "2024-09-24T18:54:24.007740Z", "iso8601": "2024-09-24T18:54:24Z", "iso8601_basic": "20240924T145424007740", "iso8601_basic_short": "20240924T145424", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "gather_subset": ["all"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["all"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} <<< 13118 1727204064.01684: stdout chunk (state=3): >>># clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.path_hooks # clear sys.path_importer_cache # clear sys.meta_path # clear sys.__interactivehook__ # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib <<< 13118 1727204064.01749: stdout chunk (state=3): >>># cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing encodings.latin_1 # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing _locale # cleanup[2] removing _bootlocale # destroy _bootlocale # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing sre_constants # destroy sre_constants # cleanup[2] removing sre_parse # cleanup[2] removing sre_compile # cleanup[2] removing _heapq # cleanup[2] removing heapq # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing collections.abc # cleanup[2] removing contextlib # cleanup[2] removing typing # destroy typing # cleanup[2] removing importlib.abc # cleanup[2] removing importlib.util # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing threading # cleanup[2] removing _bz2 # destroy _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing pwd # cleanup[2] removing grp # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing tempfile # cleanup[2] removing zipfile # destroy zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing fcntl # cleanup[2] removing locale # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing subprocess # cleanup[2] removing platform # cleanup[2] removing shlex # cleanup[2] removing token # destroy token # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc <<< 13118 1727204064.01832: stdout chunk (state=3): >>># cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # cleanup[2] removing ansible.module_utils.facts.namespace # cleanup[2] removing ansible.module_utils.compat.typing # cleanup[2] removing multiprocessing.process # cleanup[2] removing _compat_pickle # cleanup[2] removing _pickle # cleanup[2] removing pickle # cleanup[2] removing multiprocessing.reduction # cleanup[2] removing multiprocessing.context # cleanup[2] removing __mp_main__ # destroy __main__ # cleanup[2] removing multiprocessing # cleanup[2] removing _queue # cleanup[2] removing queue # cleanup[2] removing multiprocessing.util # cleanup[2] removing _multiprocessing # cleanup[2] removing multiprocessing.connection # cleanup[2] removing multiprocessing.pool # cleanup[2] removing ansible.module_utils.facts.timeout # cleanup[2] removing ansible.module_utils.facts.collector # cleanup[2] removing ansible.module_utils.facts.other # cleanup[2] removing ansible.module_utils.facts.other.facter # cleanup[2] removing ansible.module_utils.facts.other.ohai # cleanup[2] removing ansible.module_utils.facts.system # cleanup[2] removing ansible.module_utils.facts.system.apparmor # cleanup[2] removing ansible.module_utils.facts.system.caps # cleanup[2] removing ansible.module_utils.facts.system.chroot # cleanup[2] removing ansible.module_utils.facts.utils # cleanup[2] removing ansible.module_utils.facts.system.cmdline # cleanup[2] removing ansible.module_utils.facts.system.distribution # cleanup[2] removing ansible.module_utils.compat.datetime # destroy ansible.module_utils.compat.datetime # cleanup[2] removing ansible.module_utils.facts.system.date_time # cleanup[2] removing ansible.module_utils.facts.system.env # cleanup[2] removing ansible.module_utils.facts.system.dns # cleanup[2] removing ansible.module_utils.facts.system.fips # cleanup[2] removing ansible.module_utils.facts.system.loadavg # cleanup[2] removing glob # cleanup[2] removing configparser # cleanup[2] removing ansible.module_utils.facts.system.local # cleanup[2] removing ansible.module_utils.facts.system.lsb # cleanup[2] removing ansible.module_utils.facts.system.pkg_mgr # cleanup[2] removing ansible.module_utils.facts.system.platform # cleanup[2] removing _ssl # cleanup[2] removing ssl # destroy ssl # cleanup[2] removing ansible.module_utils.facts.system.python # cleanup[2] removing ansible.module_utils.facts.system.selinux # cleanup[2] removing ansible.module_utils.compat.version # destroy ansible.module_utils.compat.version # cleanup[2] removing ansible.module_utils.facts.system.service_mgr # cleanup[2] removing ansible.module_utils.facts.system.ssh_pub_keys # cleanup[2] removing termios # cleanup[2] removing getpass # cleanup[2] removing ansible.module_utils.facts.system.user # cleanup[2] removing ansible.module_utils.facts.hardware # cleanup[2] removing ansible.module_utils.facts.hardware.base # cleanup[2] removing ansible.module_utils.facts.hardware.aix # cleanup[2] removing ansible.module_utils.facts.sysctl # cleanup[2] removing ansible.module_utils.facts.hardware.darwin # cleanup[2] removing ansible.module_utils.facts.hardware.freebsd # cleanup[2] removing ansible.module_utils.facts.hardware.dragonfly # cleanup[2] removing ansible.module_utils.facts.hardware.hpux # cleanup[2] removing ansible.module_utils.facts.hardware.linux # cleanup[2] removing ansible.module_utils.facts.hardware.hurd # cleanup[2] removing ansible.module_utils.facts.hardware.netbsd # cleanup[2] removing ansible.module_utils.facts.hardware.openbsd # cleanup[2] removing ansible.module_utils.facts.hardware.sunos # cleanup[2] removing ansible.module_utils.facts.network # cleanup[2] removing ansible.module_utils.facts.network.base # cleanup[2] removing ansible.module_utils.facts.network.generic_bsd # cleanup[2] removing ansible.module_utils.facts.network.aix # cleanup[2] removing ansible.module_utils.facts.network.darwin # cleanup[2] removing ansible.module_utils.facts.network.dragonfly # cleanup[2] removing ansible.module_utils.facts.network.fc_wwn # cleanup[2] removing ansible.module_utils.facts.network.freebsd # cleanup[2] removing ansible.module_utils.facts.network.hpux # cleanup[2] removing ansible.module_utils.facts.network.hurd # cleanup[2] removing ansible.module_utils.facts.network.linux # cleanup[2] removing ansible.module_utils.facts.network.iscsi # cleanup[2] removing ansible.module_utils.facts.network.nvme # cleanup[2] removing ansible.module_utils.facts.network.netbsd # cleanup[2] removing ansible.module_utils.facts.network.openbsd # cleanup[2] removing ansible.module_utils.facts.network.sunos # cleanup[2] removing ansible.module_utils.facts.virtual # cleanup[2] removing ansible.module_utils.facts.virtual.base # cleanup[2] removing ansible.module_utils.facts.virtual.sysctl # cleanup[2] removing ansible.module_utils.facts.virtual.freebsd # cleanup[2] removing ansible.module_utils.facts.virtual.dragonfly # cleanup[2] removing ansible.module_utils.facts.virtual.hpux # cleanup[2] removing ansible.module_utils.facts.virtual.linux # cleanup[2] removing ansible.module_utils.facts.virtual.netbsd # cleanup[2] removing ansible.module_utils.facts.virtual.openbsd # cleanup[2] removing ansible.module_utils.facts.virtual.sunos # cleanup[2] removing ansible.module_utils.facts.default_collectors # cleanup[2] removing ansible.module_utils.facts.ansible_collector # cleanup[2] removing ansible.module_utils.facts.compat # cleanup[2] removing ansible.module_utils.facts # destroy ansible.module_utils.facts # destroy ansible.module_utils.facts.namespace # destroy ansible.module_utils.facts.other # destroy ansible.module_utils.facts.other.facter # destroy ansible.module_utils.facts.other.ohai # destroy ansible.module_utils.facts.system # destroy ansible.module_utils.facts.system.apparmor # destroy ansible.module_utils.facts.system.caps # destroy ansible.module_utils.facts.system.chroot # destroy ansible.module_utils.facts.system.cmdline # destroy ansible.module_utils.facts.system.distribution # destroy ansible.module_utils.facts.system.date_time # destroy ansible.module_utils.facts.system.env # destroy ansible.module_utils.facts.system.dns # destroy ansible.module_utils.facts.system.fips # destroy ansible.module_utils.facts.system.loadavg # destroy ansible.module_utils.facts.system.local # destroy ansible.module_utils.facts.system.lsb # destroy ansible.module_utils.facts.system.pkg_mgr # destroy ansible.module_utils.facts.system.platform # destroy ansible.module_utils.facts.system.python # destroy ansible.module_utils.facts.system.selinux # destroy ansible.module_utils.facts.system.service_mgr # destroy ansible.module_utils.facts.system.ssh_pub_keys # destroy ansible.module_utils.facts.system.user # destroy ansible.module_utils.facts.utils # destroy ansible.module_utils.facts.hardware # destroy ansible.module_utils.facts.hardware.base # destroy ansible.module_utils.facts.hardware.aix # destroy ansible.module_utils.facts.hardware.darwin # destroy ansible.module_utils.facts.hardware.freebsd # destroy ansible.module_utils.facts.hardware.dragonfly # destroy ansible.module_utils.facts.hardware.hpux # destroy ansible.module_utils.facts.hardware.linux # destroy ansible.module_utils.facts.hardware.hurd # destroy ansible.module_utils.facts.hardware.netbsd # destroy ansible.module_utils.facts.hardware.openbsd # destroy ansible.module_utils.facts.hardware.sunos # destroy ansible.module_utils.facts.sysctl # destroy ansible.module_utils.facts.network # destroy ansible.module_utils.facts.network.base # destroy ansible.module_utils.facts.network.generic_bsd # destroy ansible.module_utils.facts.network.aix # destroy ansible.module_utils.facts.network.darwin # destroy ansible.module_utils.facts.network.dragonfly # destroy ansible.module_utils.facts.network.fc_wwn # destroy ansible.module_utils.facts.network.freebsd # destroy ansible.module_utils.facts.network.hpux # destroy ansible.module_utils.facts.network.hurd # destroy ansible.module_utils.facts.network.linux # destroy ansible.module_utils.facts.network.iscsi # destroy ansible.module_utils.facts.network.nvme # destroy ansible.module_utils.facts.network.netbsd # destroy ansible.module_utils.facts.network.openbsd # destroy ansible.module_utils.facts.network.sunos # destroy ansible.module_utils.facts.virtual # destroy ansible.module_utils.facts.virtual.base # destroy ansible.module_utils.facts.virtual.sysctl # destroy ansible.module_utils.facts.virtual.freebsd # destroy ansible.module_utils.facts.virtual.dragonfly # destroy ansible.module_utils.facts.virtual.hpux # destroy ansible.module_utils.facts.virtual.linux <<< 13118 1727204064.01840: stdout chunk (state=3): >>># destroy ansible.module_utils.facts.virtual.netbsd # destroy ansible.module_utils.facts.virtual.openbsd # destroy ansible.module_utils.facts.virtual.sunos # destroy ansible.module_utils.facts.compat # cleanup[2] removing gc # cleanup[2] removing unicodedata # cleanup[2] removing stringprep # cleanup[2] removing encodings.idna # cleanup[2] removing multiprocessing.queues # cleanup[2] removing multiprocessing.synchronize # cleanup[2] removing multiprocessing.dummy.connection # cleanup[2] removing multiprocessing.dummy <<< 13118 1727204064.02098: stdout chunk (state=3): >>># destroy _sitebuiltins <<< 13118 1727204064.02122: stdout chunk (state=3): >>># destroy importlib.util # destroy importlib.abc # destroy importlib.machinery <<< 13118 1727204064.02151: stdout chunk (state=3): >>># destroy zipimport # destroy _compression # destroy binascii # destroy importlib # destroy bz2 # destroy lzma <<< 13118 1727204064.02200: stdout chunk (state=3): >>># destroy __main__ # destroy locale # destroy systemd.journal # destroy systemd.daemon # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy encodings <<< 13118 1727204064.02223: stdout chunk (state=3): >>># destroy syslog # destroy uuid <<< 13118 1727204064.02266: stdout chunk (state=3): >>># destroy selinux # destroy distro # destroy logging # destroy argparse <<< 13118 1727204064.02333: stdout chunk (state=3): >>># destroy ansible.module_utils.facts.default_collectors # destroy ansible.module_utils.facts.ansible_collector # destroy multiprocessing # destroy multiprocessing.queues # destroy multiprocessing.synchronize # destroy multiprocessing.dummy # destroy multiprocessing.pool # destroy pickle # destroy _compat_pickle <<< 13118 1727204064.02370: stdout chunk (state=3): >>># destroy queue # destroy multiprocessing.reduction <<< 13118 1727204064.02420: stdout chunk (state=3): >>># destroy shlex # destroy datetime # destroy base64 <<< 13118 1727204064.02475: stdout chunk (state=3): >>># destroy ansible.module_utils.compat.selinux # destroy getpass # destroy json # destroy socket # destroy struct # destroy glob # destroy ansible.module_utils.compat.typing # destroy ansible.module_utils.facts.timeout # destroy ansible.module_utils.facts.collector # destroy multiprocessing.connection # destroy tempfile # destroy multiprocessing.context # destroy multiprocessing.process # destroy multiprocessing.util <<< 13118 1727204064.02488: stdout chunk (state=3): >>># destroy array # destroy multiprocessing.dummy.connection <<< 13118 1727204064.02552: stdout chunk (state=3): >>># cleanup[3] wiping encodings.idna # destroy stringprep # cleanup[3] wiping unicodedata # cleanup[3] wiping gc # cleanup[3] wiping termios # cleanup[3] wiping _ssl # cleanup[3] wiping configparser # cleanup[3] wiping _multiprocessing # cleanup[3] wiping _queue # cleanup[3] wiping _pickle # cleanup[3] wiping selinux._selinux <<< 13118 1727204064.02650: stdout chunk (state=3): >>># cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # destroy configparser # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # cleanup[3] wiping tokenize # cleanup[3] wiping platform # destroy subprocess # cleanup[3] wiping selectors # cleanup[3] wiping select # cleanup[3] wiping _posixsubprocess # cleanup[3] wiping signal # cleanup[3] wiping fcntl # cleanup[3] wiping atexit # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping _blake2 # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping shutil # destroy fnmatch # cleanup[3] wiping grp # cleanup[3] wiping pwd # cleanup[3] wiping _lzma # cleanup[3] wiping threading # cleanup[3] wiping zlib # cleanup[3] wiping errno # cleanup[3] wiping weakref # cleanup[3] wiping contextlib # cleanup[3] wiping collections.abc # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re <<< 13118 1727204064.02702: stdout chunk (state=3): >>># destroy enum # destroy sre_compile # destroy copyreg # cleanup[3] wiping functools # cleanup[3] wiping _functools # destroy _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy heapq # destroy collections.abc # cleanup[3] wiping _collections # destroy _collections # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping itertools # cleanup[3] wiping _heapq # cleanup[3] wiping sre_parse # cleanup[3] wiping _sre # cleanup[3] wiping types # cleanup[3] wiping _locale # destroy _locale # cleanup[3] wiping os # cleanup[3] wiping os.path # destroy genericpath <<< 13118 1727204064.02745: stdout chunk (state=3): >>># cleanup[3] wiping posixpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.latin_1 # cleanup[3] wiping _signal # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins <<< 13118 1727204064.02772: stdout chunk (state=3): >>># destroy unicodedata # destroy gc # destroy termios # destroy _ssl # destroy _multiprocessing # destroy _queue # destroy _pickle # destroy systemd._daemon # destroy _socket # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy fcntl # destroy _blake2 # destroy _lzma # destroy zlib # destroy _signal <<< 13118 1727204064.02940: stdout chunk (state=3): >>># destroy platform # destroy _uuid # destroy _sre # destroy sre_parse # destroy tokenize <<< 13118 1727204064.02983: stdout chunk (state=3): >>># destroy _heapq # destroy posixpath # destroy stat <<< 13118 1727204064.02998: stdout chunk (state=3): >>># destroy ansible.module_utils.six.moves.urllib # destroy errno # destroy signal # destroy contextlib # destroy pwd # destroy grp # destroy _posixsubprocess # destroy selectors <<< 13118 1727204064.03011: stdout chunk (state=3): >>># destroy select # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy itertools # destroy operator # destroy ansible.module_utils.six.moves # destroy _operator # destroy _frozen_importlib_external # destroy _imp # destroy io # destroy marshal <<< 13118 1727204064.03048: stdout chunk (state=3): >>># destroy _frozen_importlib # clear sys.audit hooks <<< 13118 1727204064.03433: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204064.03443: stdout chunk (state=3): >>><<< 13118 1727204064.03457: stderr chunk (state=3): >>><<< 13118 1727204064.03621: _low_level_execute_command() done: rc=0, stdout=import _frozen_importlib # frozen import _imp # builtin import '_thread' # import '_warnings' # import '_weakref' # import '_io' # import 'marshal' # import 'posix' # import '_frozen_importlib_external' # # installing zipimport hook import 'time' # import 'zipimport' # # installed zipimport hook # /usr/lib64/python3.9/encodings/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/encodings/__init__.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/codecs.cpython-39.pyc matches /usr/lib64/python3.9/codecs.py # code object from '/usr/lib64/python3.9/__pycache__/codecs.cpython-39.pyc' import '_codecs' # import 'codecs' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a843dc0> # /usr/lib64/python3.9/encodings/__pycache__/aliases.cpython-39.pyc matches /usr/lib64/python3.9/encodings/aliases.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/aliases.cpython-39.pyc' import 'encodings.aliases' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5d83a0> import 'encodings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a843b20> # /usr/lib64/python3.9/encodings/__pycache__/utf_8.cpython-39.pyc matches /usr/lib64/python3.9/encodings/utf_8.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/utf_8.cpython-39.pyc' import 'encodings.utf_8' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a843ac0> import '_signal' # # /usr/lib64/python3.9/encodings/__pycache__/latin_1.cpython-39.pyc matches /usr/lib64/python3.9/encodings/latin_1.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/latin_1.cpython-39.pyc' import 'encodings.latin_1' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5d8490> # /usr/lib64/python3.9/__pycache__/io.cpython-39.pyc matches /usr/lib64/python3.9/io.py # code object from '/usr/lib64/python3.9/__pycache__/io.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/abc.py # code object from '/usr/lib64/python3.9/__pycache__/abc.cpython-39.pyc' import '_abc' # import 'abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5d8940> import 'io' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5d8670> # /usr/lib64/python3.9/__pycache__/site.cpython-39.pyc matches /usr/lib64/python3.9/site.py # code object from '/usr/lib64/python3.9/__pycache__/site.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/os.cpython-39.pyc matches /usr/lib64/python3.9/os.py # code object from '/usr/lib64/python3.9/__pycache__/os.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/stat.cpython-39.pyc matches /usr/lib64/python3.9/stat.py # code object from '/usr/lib64/python3.9/__pycache__/stat.cpython-39.pyc' import '_stat' # import 'stat' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a58f190> # /usr/lib64/python3.9/__pycache__/_collections_abc.cpython-39.pyc matches /usr/lib64/python3.9/_collections_abc.py # code object from '/usr/lib64/python3.9/__pycache__/_collections_abc.cpython-39.pyc' import '_collections_abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a58f220> # /usr/lib64/python3.9/__pycache__/posixpath.cpython-39.pyc matches /usr/lib64/python3.9/posixpath.py # code object from '/usr/lib64/python3.9/__pycache__/posixpath.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/genericpath.cpython-39.pyc matches /usr/lib64/python3.9/genericpath.py # code object from '/usr/lib64/python3.9/__pycache__/genericpath.cpython-39.pyc' import 'genericpath' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5b2850> import 'posixpath' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a58f940> import 'os' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5f0880> # /usr/lib64/python3.9/__pycache__/_sitebuiltins.cpython-39.pyc matches /usr/lib64/python3.9/_sitebuiltins.py # code object from '/usr/lib64/python3.9/__pycache__/_sitebuiltins.cpython-39.pyc' import '_sitebuiltins' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a588d90> # /usr/lib64/python3.9/__pycache__/_bootlocale.cpython-39.pyc matches /usr/lib64/python3.9/_bootlocale.py # code object from '/usr/lib64/python3.9/__pycache__/_bootlocale.cpython-39.pyc' import '_locale' # import '_bootlocale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5b2d90> import 'site' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5d8970> Python 3.9.19 (main, Aug 23 2024, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-2)] on linux Type "help", "copyright", "credits" or "license" for more information. # /usr/lib64/python3.9/__pycache__/base64.cpython-39.pyc matches /usr/lib64/python3.9/base64.py # code object from '/usr/lib64/python3.9/__pycache__/base64.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/re.cpython-39.pyc matches /usr/lib64/python3.9/re.py # code object from '/usr/lib64/python3.9/__pycache__/re.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/enum.cpython-39.pyc matches /usr/lib64/python3.9/enum.py # code object from '/usr/lib64/python3.9/__pycache__/enum.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/types.cpython-39.pyc matches /usr/lib64/python3.9/types.py # code object from '/usr/lib64/python3.9/__pycache__/types.cpython-39.pyc' import 'types' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a52ef10> import 'enum' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5340a0> # /usr/lib64/python3.9/__pycache__/sre_compile.cpython-39.pyc matches /usr/lib64/python3.9/sre_compile.py # code object from '/usr/lib64/python3.9/__pycache__/sre_compile.cpython-39.pyc' import '_sre' # # /usr/lib64/python3.9/__pycache__/sre_parse.cpython-39.pyc matches /usr/lib64/python3.9/sre_parse.py # code object from '/usr/lib64/python3.9/__pycache__/sre_parse.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/sre_constants.cpython-39.pyc matches /usr/lib64/python3.9/sre_constants.py # code object from '/usr/lib64/python3.9/__pycache__/sre_constants.cpython-39.pyc' import 'sre_constants' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5275b0> import 'sre_parse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a52f6a0> import 'sre_compile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a52e3d0> # /usr/lib64/python3.9/__pycache__/functools.cpython-39.pyc matches /usr/lib64/python3.9/functools.py # code object from '/usr/lib64/python3.9/__pycache__/functools.cpython-39.pyc' # /usr/lib64/python3.9/collections/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/collections/__init__.py # code object from '/usr/lib64/python3.9/collections/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/heapq.cpython-39.pyc matches /usr/lib64/python3.9/heapq.py # code object from '/usr/lib64/python3.9/__pycache__/heapq.cpython-39.pyc' # extension module '_heapq' loaded from '/usr/lib64/python3.9/lib-dynload/_heapq.cpython-39-x86_64-linux-gnu.so' # extension module '_heapq' executed from '/usr/lib64/python3.9/lib-dynload/_heapq.cpython-39-x86_64-linux-gnu.so' import '_heapq' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a415e50> import 'heapq' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a415940> import 'itertools' # # /usr/lib64/python3.9/__pycache__/keyword.cpython-39.pyc matches /usr/lib64/python3.9/keyword.py # code object from '/usr/lib64/python3.9/__pycache__/keyword.cpython-39.pyc' import 'keyword' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a415f40> # /usr/lib64/python3.9/__pycache__/operator.cpython-39.pyc matches /usr/lib64/python3.9/operator.py # code object from '/usr/lib64/python3.9/__pycache__/operator.cpython-39.pyc' import '_operator' # import 'operator' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a415d90> # /usr/lib64/python3.9/__pycache__/reprlib.cpython-39.pyc matches /usr/lib64/python3.9/reprlib.py # code object from '/usr/lib64/python3.9/__pycache__/reprlib.cpython-39.pyc' import 'reprlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a426100> import '_collections' # import 'collections' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a509dc0> import '_functools' # import 'functools' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5026a0> # /usr/lib64/python3.9/__pycache__/copyreg.cpython-39.pyc matches /usr/lib64/python3.9/copyreg.py # code object from '/usr/lib64/python3.9/__pycache__/copyreg.cpython-39.pyc' import 'copyreg' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a515700> import 're' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a535eb0> # /usr/lib64/python3.9/__pycache__/struct.cpython-39.pyc matches /usr/lib64/python3.9/struct.py # code object from '/usr/lib64/python3.9/__pycache__/struct.cpython-39.pyc' # extension module '_struct' loaded from '/usr/lib64/python3.9/lib-dynload/_struct.cpython-39-x86_64-linux-gnu.so' # extension module '_struct' executed from '/usr/lib64/python3.9/lib-dynload/_struct.cpython-39-x86_64-linux-gnu.so' import '_struct' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a426d00> import 'struct' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a5092e0> # extension module 'binascii' loaded from '/usr/lib64/python3.9/lib-dynload/binascii.cpython-39-x86_64-linux-gnu.so' # extension module 'binascii' executed from '/usr/lib64/python3.9/lib-dynload/binascii.cpython-39-x86_64-linux-gnu.so' import 'binascii' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a515310> import 'base64' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a53ba60> # /usr/lib64/python3.9/__pycache__/runpy.cpython-39.pyc matches /usr/lib64/python3.9/runpy.py # code object from '/usr/lib64/python3.9/__pycache__/runpy.cpython-39.pyc' # /usr/lib64/python3.9/importlib/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/importlib/__init__.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/warnings.cpython-39.pyc matches /usr/lib64/python3.9/warnings.py # code object from '/usr/lib64/python3.9/__pycache__/warnings.cpython-39.pyc' import 'warnings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a426ee0> import 'importlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a426e20> # /usr/lib64/python3.9/importlib/__pycache__/machinery.cpython-39.pyc matches /usr/lib64/python3.9/importlib/machinery.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/machinery.cpython-39.pyc' import 'importlib.machinery' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a426d90> # /usr/lib64/python3.9/importlib/__pycache__/util.cpython-39.pyc matches /usr/lib64/python3.9/importlib/util.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/util.cpython-39.pyc' # /usr/lib64/python3.9/importlib/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/importlib/abc.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/abc.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/typing.cpython-39.pyc matches /usr/lib64/python3.9/typing.py # code object from '/usr/lib64/python3.9/__pycache__/typing.cpython-39.pyc' # /usr/lib64/python3.9/collections/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/collections/abc.py # code object from '/usr/lib64/python3.9/collections/__pycache__/abc.cpython-39.pyc' import 'collections.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a3f9400> # /usr/lib64/python3.9/__pycache__/contextlib.cpython-39.pyc matches /usr/lib64/python3.9/contextlib.py # code object from '/usr/lib64/python3.9/__pycache__/contextlib.cpython-39.pyc' import 'contextlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a3f94f0> import 'typing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a42ef70> import 'importlib.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a428ac0> import 'importlib.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a428490> # /usr/lib64/python3.9/__pycache__/pkgutil.cpython-39.pyc matches /usr/lib64/python3.9/pkgutil.py # code object from '/usr/lib64/python3.9/__pycache__/pkgutil.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/weakref.cpython-39.pyc matches /usr/lib64/python3.9/weakref.py # code object from '/usr/lib64/python3.9/__pycache__/weakref.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/_weakrefset.cpython-39.pyc matches /usr/lib64/python3.9/_weakrefset.py # code object from '/usr/lib64/python3.9/__pycache__/_weakrefset.cpython-39.pyc' import '_weakrefset' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a32d250> import 'weakref' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a3e4550> import 'pkgutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a428f40> import 'runpy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a53b0d0> # /usr/lib64/python3.9/__pycache__/shutil.cpython-39.pyc matches /usr/lib64/python3.9/shutil.py # code object from '/usr/lib64/python3.9/__pycache__/shutil.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/fnmatch.cpython-39.pyc matches /usr/lib64/python3.9/fnmatch.py # code object from '/usr/lib64/python3.9/__pycache__/fnmatch.cpython-39.pyc' import 'fnmatch' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a33fb80> import 'errno' # # extension module 'zlib' loaded from '/usr/lib64/python3.9/lib-dynload/zlib.cpython-39-x86_64-linux-gnu.so' # extension module 'zlib' executed from '/usr/lib64/python3.9/lib-dynload/zlib.cpython-39-x86_64-linux-gnu.so' import 'zlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a33feb0> # /usr/lib64/python3.9/__pycache__/bz2.cpython-39.pyc matches /usr/lib64/python3.9/bz2.py # code object from '/usr/lib64/python3.9/__pycache__/bz2.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/_compression.cpython-39.pyc matches /usr/lib64/python3.9/_compression.py # code object from '/usr/lib64/python3.9/__pycache__/_compression.cpython-39.pyc' import '_compression' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a3507c0> # /usr/lib64/python3.9/__pycache__/threading.cpython-39.pyc matches /usr/lib64/python3.9/threading.py # code object from '/usr/lib64/python3.9/__pycache__/threading.cpython-39.pyc' import 'threading' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a350d00> # extension module '_bz2' loaded from '/usr/lib64/python3.9/lib-dynload/_bz2.cpython-39-x86_64-linux-gnu.so' # extension module '_bz2' executed from '/usr/lib64/python3.9/lib-dynload/_bz2.cpython-39-x86_64-linux-gnu.so' import '_bz2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a2de430> import 'bz2' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a33ffa0> # /usr/lib64/python3.9/__pycache__/lzma.cpython-39.pyc matches /usr/lib64/python3.9/lzma.py # code object from '/usr/lib64/python3.9/__pycache__/lzma.cpython-39.pyc' # extension module '_lzma' loaded from '/usr/lib64/python3.9/lib-dynload/_lzma.cpython-39-x86_64-linux-gnu.so' # extension module '_lzma' executed from '/usr/lib64/python3.9/lib-dynload/_lzma.cpython-39-x86_64-linux-gnu.so' import '_lzma' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a2ef310> import 'lzma' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a350640> import 'pwd' # # extension module 'grp' loaded from '/usr/lib64/python3.9/lib-dynload/grp.cpython-39-x86_64-linux-gnu.so' # extension module 'grp' executed from '/usr/lib64/python3.9/lib-dynload/grp.cpython-39-x86_64-linux-gnu.so' import 'grp' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a2ef3d0> import 'shutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a426a60> # /usr/lib64/python3.9/__pycache__/tempfile.cpython-39.pyc matches /usr/lib64/python3.9/tempfile.py # code object from '/usr/lib64/python3.9/__pycache__/tempfile.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/random.cpython-39.pyc matches /usr/lib64/python3.9/random.py # code object from '/usr/lib64/python3.9/__pycache__/random.cpython-39.pyc' # extension module 'math' loaded from '/usr/lib64/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so' # extension module 'math' executed from '/usr/lib64/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so' import 'math' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a30b730> # /usr/lib64/python3.9/__pycache__/bisect.cpython-39.pyc matches /usr/lib64/python3.9/bisect.py # code object from '/usr/lib64/python3.9/__pycache__/bisect.cpython-39.pyc' # extension module '_bisect' loaded from '/usr/lib64/python3.9/lib-dynload/_bisect.cpython-39-x86_64-linux-gnu.so' # extension module '_bisect' executed from '/usr/lib64/python3.9/lib-dynload/_bisect.cpython-39-x86_64-linux-gnu.so' import '_bisect' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a30ba00> import 'bisect' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a30b7f0> # extension module '_random' loaded from '/usr/lib64/python3.9/lib-dynload/_random.cpython-39-x86_64-linux-gnu.so' # extension module '_random' executed from '/usr/lib64/python3.9/lib-dynload/_random.cpython-39-x86_64-linux-gnu.so' import '_random' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a30b8e0> # /usr/lib64/python3.9/__pycache__/hashlib.cpython-39.pyc matches /usr/lib64/python3.9/hashlib.py # code object from '/usr/lib64/python3.9/__pycache__/hashlib.cpython-39.pyc' # extension module '_hashlib' loaded from '/usr/lib64/python3.9/lib-dynload/_hashlib.cpython-39-x86_64-linux-gnu.so' # extension module '_hashlib' executed from '/usr/lib64/python3.9/lib-dynload/_hashlib.cpython-39-x86_64-linux-gnu.so' import '_hashlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a30bd30> # extension module '_blake2' loaded from '/usr/lib64/python3.9/lib-dynload/_blake2.cpython-39-x86_64-linux-gnu.so' # extension module '_blake2' executed from '/usr/lib64/python3.9/lib-dynload/_blake2.cpython-39-x86_64-linux-gnu.so' import '_blake2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a315280> import 'hashlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a30b970> import 'random' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a2feac0> import 'tempfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a426640> # /usr/lib64/python3.9/__pycache__/zipfile.cpython-39.pyc matches /usr/lib64/python3.9/zipfile.py # code object from '/usr/lib64/python3.9/__pycache__/zipfile.cpython-39.pyc' import 'zipfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a30bb20> # code object from '/usr/lib64/python3.9/encodings/cp437.pyc' import 'encodings.cp437' # <_frozen_importlib_external.SourcelessFileLoader object at 0x7f111a232700> # zipimport: found 103 names in '/tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip' # zipimport: zlib available # zipimport: zlib available import ansible # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/__init__.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/__future__.cpython-39.pyc matches /usr/lib64/python3.9/__future__.py # code object from '/usr/lib64/python3.9/__pycache__/__future__.cpython-39.pyc' import '__future__' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a171850> # /usr/lib64/python3.9/json/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/json/__init__.py # code object from '/usr/lib64/python3.9/json/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/json/__pycache__/decoder.cpython-39.pyc matches /usr/lib64/python3.9/json/decoder.py # code object from '/usr/lib64/python3.9/json/__pycache__/decoder.cpython-39.pyc' # /usr/lib64/python3.9/json/__pycache__/scanner.cpython-39.pyc matches /usr/lib64/python3.9/json/scanner.py # code object from '/usr/lib64/python3.9/json/__pycache__/scanner.cpython-39.pyc' # extension module '_json' loaded from '/usr/lib64/python3.9/lib-dynload/_json.cpython-39-x86_64-linux-gnu.so' # extension module '_json' executed from '/usr/lib64/python3.9/lib-dynload/_json.cpython-39-x86_64-linux-gnu.so' import '_json' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a171160> import 'json.scanner' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a171280> import 'json.decoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a171fa0> # /usr/lib64/python3.9/json/__pycache__/encoder.cpython-39.pyc matches /usr/lib64/python3.9/json/encoder.py # code object from '/usr/lib64/python3.9/json/__pycache__/encoder.cpython-39.pyc' import 'json.encoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1714f0> import 'json' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a171dc0> import 'atexit' # # extension module 'fcntl' loaded from '/usr/lib64/python3.9/lib-dynload/fcntl.cpython-39-x86_64-linux-gnu.so' # extension module 'fcntl' executed from '/usr/lib64/python3.9/lib-dynload/fcntl.cpython-39-x86_64-linux-gnu.so' import 'fcntl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a171580> # /usr/lib64/python3.9/__pycache__/locale.cpython-39.pyc matches /usr/lib64/python3.9/locale.py # code object from '/usr/lib64/python3.9/__pycache__/locale.cpython-39.pyc' import 'locale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a171100> # /usr/lib64/python3.9/__pycache__/platform.cpython-39.pyc matches /usr/lib64/python3.9/platform.py # code object from '/usr/lib64/python3.9/__pycache__/platform.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/subprocess.cpython-39.pyc matches /usr/lib64/python3.9/subprocess.py # code object from '/usr/lib64/python3.9/__pycache__/subprocess.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/signal.cpython-39.pyc matches /usr/lib64/python3.9/signal.py # code object from '/usr/lib64/python3.9/__pycache__/signal.cpython-39.pyc' import 'signal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1470a0> # extension module '_posixsubprocess' loaded from '/usr/lib64/python3.9/lib-dynload/_posixsubprocess.cpython-39-x86_64-linux-gnu.so' # extension module '_posixsubprocess' executed from '/usr/lib64/python3.9/lib-dynload/_posixsubprocess.cpython-39-x86_64-linux-gnu.so' import '_posixsubprocess' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b08370> # extension module 'select' loaded from '/usr/lib64/python3.9/lib-dynload/select.cpython-39-x86_64-linux-gnu.so' # extension module 'select' executed from '/usr/lib64/python3.9/lib-dynload/select.cpython-39-x86_64-linux-gnu.so' import 'select' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b08070> # /usr/lib64/python3.9/__pycache__/selectors.cpython-39.pyc matches /usr/lib64/python3.9/selectors.py # code object from '/usr/lib64/python3.9/__pycache__/selectors.cpython-39.pyc' import 'selectors' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119b08cd0> import 'subprocess' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a159dc0> import 'platform' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1593a0> # /usr/lib64/python3.9/__pycache__/shlex.cpython-39.pyc matches /usr/lib64/python3.9/shlex.py # code object from '/usr/lib64/python3.9/__pycache__/shlex.cpython-39.pyc' import 'shlex' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a159f40> # /usr/lib64/python3.9/__pycache__/traceback.cpython-39.pyc matches /usr/lib64/python3.9/traceback.py # code object from '/usr/lib64/python3.9/__pycache__/traceback.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/linecache.cpython-39.pyc matches /usr/lib64/python3.9/linecache.py # code object from '/usr/lib64/python3.9/__pycache__/linecache.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/tokenize.cpython-39.pyc matches /usr/lib64/python3.9/tokenize.py # code object from '/usr/lib64/python3.9/__pycache__/tokenize.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/token.cpython-39.pyc matches /usr/lib64/python3.9/token.py # code object from '/usr/lib64/python3.9/__pycache__/token.cpython-39.pyc' import 'token' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1a6f40> import 'tokenize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a173d60> import 'linecache' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a173430> import 'traceback' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119be1af0> # extension module 'syslog' loaded from '/usr/lib64/python3.9/lib-dynload/syslog.cpython-39-x86_64-linux-gnu.so' # extension module 'syslog' executed from '/usr/lib64/python3.9/lib-dynload/syslog.cpython-39-x86_64-linux-gnu.so' import 'syslog' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a173550> # /usr/lib64/python3.9/site-packages/systemd/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/__init__.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/__init__.cpython-39.pyc' import 'systemd' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a173580> # /usr/lib64/python3.9/site-packages/systemd/__pycache__/journal.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/journal.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/journal.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/datetime.cpython-39.pyc matches /usr/lib64/python3.9/datetime.py # code object from '/usr/lib64/python3.9/__pycache__/datetime.cpython-39.pyc' # extension module '_datetime' loaded from '/usr/lib64/python3.9/lib-dynload/_datetime.cpython-39-x86_64-linux-gnu.so' # extension module '_datetime' executed from '/usr/lib64/python3.9/lib-dynload/_datetime.cpython-39-x86_64-linux-gnu.so' import '_datetime' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b76fa0> import 'datetime' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1b8280> # /usr/lib64/python3.9/__pycache__/uuid.cpython-39.pyc matches /usr/lib64/python3.9/uuid.py # code object from '/usr/lib64/python3.9/__pycache__/uuid.cpython-39.pyc' # extension module '_uuid' loaded from '/usr/lib64/python3.9/lib-dynload/_uuid.cpython-39-x86_64-linux-gnu.so' # extension module '_uuid' executed from '/usr/lib64/python3.9/lib-dynload/_uuid.cpython-39-x86_64-linux-gnu.so' import '_uuid' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b73820> import 'uuid' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1b8400> # /usr/lib64/python3.9/logging/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/logging/__init__.py # code object from '/usr/lib64/python3.9/logging/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/string.cpython-39.pyc matches /usr/lib64/python3.9/string.py # code object from '/usr/lib64/python3.9/__pycache__/string.cpython-39.pyc' import '_string' # import 'string' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1b8c40> import 'logging' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119b737c0> # extension module 'systemd._journal' loaded from '/usr/lib64/python3.9/site-packages/systemd/_journal.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._journal' executed from '/usr/lib64/python3.9/site-packages/systemd/_journal.cpython-39-x86_64-linux-gnu.so' import 'systemd._journal' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a1511c0> # extension module 'systemd._reader' loaded from '/usr/lib64/python3.9/site-packages/systemd/_reader.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._reader' executed from '/usr/lib64/python3.9/site-packages/systemd/_reader.cpython-39-x86_64-linux-gnu.so' import 'systemd._reader' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a1b89d0> # extension module 'systemd.id128' loaded from '/usr/lib64/python3.9/site-packages/systemd/id128.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd.id128' executed from '/usr/lib64/python3.9/site-packages/systemd/id128.cpython-39-x86_64-linux-gnu.so' import 'systemd.id128' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111a1b8550> import 'systemd.journal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1b1940> # /usr/lib64/python3.9/site-packages/systemd/__pycache__/daemon.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/daemon.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/daemon.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/socket.cpython-39.pyc matches /usr/lib64/python3.9/socket.py # code object from '/usr/lib64/python3.9/__pycache__/socket.cpython-39.pyc' # extension module '_socket' loaded from '/usr/lib64/python3.9/lib-dynload/_socket.cpython-39-x86_64-linux-gnu.so' # extension module '_socket' executed from '/usr/lib64/python3.9/lib-dynload/_socket.cpython-39-x86_64-linux-gnu.so' import '_socket' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b68910> # extension module 'array' loaded from '/usr/lib64/python3.9/lib-dynload/array.cpython-39-x86_64-linux-gnu.so' # extension module 'array' executed from '/usr/lib64/python3.9/lib-dynload/array.cpython-39-x86_64-linux-gnu.so' import 'array' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b85dc0> import 'socket' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119b72550> # extension module 'systemd._daemon' loaded from '/usr/lib64/python3.9/site-packages/systemd/_daemon.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._daemon' executed from '/usr/lib64/python3.9/site-packages/systemd/_daemon.cpython-39-x86_64-linux-gnu.so' import 'systemd._daemon' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119b68eb0> import 'systemd.daemon' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119b72970> # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.compat # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/compat/__init__.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.text # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/text/__init__.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.six # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/six/__init__.py import 'ansible.module_utils.six.moves' # import 'ansible.module_utils.six.moves.collections_abc' # import ansible.module_utils.common.text.converters # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/text/converters.py # /usr/lib64/python3.9/ctypes/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/ctypes/__init__.py # code object from '/usr/lib64/python3.9/ctypes/__pycache__/__init__.cpython-39.pyc' # extension module '_ctypes' loaded from '/usr/lib64/python3.9/lib-dynload/_ctypes.cpython-39-x86_64-linux-gnu.so' # extension module '_ctypes' executed from '/usr/lib64/python3.9/lib-dynload/_ctypes.cpython-39-x86_64-linux-gnu.so' import '_ctypes' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119baf7f0> # /usr/lib64/python3.9/ctypes/__pycache__/_endian.cpython-39.pyc matches /usr/lib64/python3.9/ctypes/_endian.py # code object from '/usr/lib64/python3.9/ctypes/__pycache__/_endian.cpython-39.pyc' import 'ctypes._endian' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119bb48b0> import 'ctypes' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111971a940> import ansible.module_utils.compat.selinux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/compat/selinux.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils._text # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/_text.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/copy.cpython-39.pyc matches /usr/lib64/python3.9/copy.py # code object from '/usr/lib64/python3.9/__pycache__/copy.cpython-39.pyc' import 'copy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119bec730> # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.collections # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/collections.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.warnings # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/warnings.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.errors # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/errors.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.parsing # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/parsing/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.parsing.convert_bool # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/parsing/convert_bool.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/ast.cpython-39.pyc matches /usr/lib64/python3.9/ast.py # code object from '/usr/lib64/python3.9/__pycache__/ast.cpython-39.pyc' import '_ast' # import 'ast' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a1742e0> # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.text.formatters # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/text/formatters.py import ansible.module_utils.common.validation # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/validation.py import ansible.module_utils.common.parameters # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/parameters.py import ansible.module_utils.common.arg_spec # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/arg_spec.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.locale # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/locale.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/site-packages/selinux/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/selinux/__init__.py # code object from '/usr/lib64/python3.9/site-packages/selinux/__pycache__/__init__.cpython-39.pyc' # extension module 'selinux._selinux' loaded from '/usr/lib64/python3.9/site-packages/selinux/_selinux.cpython-39-x86_64-linux-gnu.so' # extension module 'selinux._selinux' executed from '/usr/lib64/python3.9/site-packages/selinux/_selinux.cpython-39-x86_64-linux-gnu.so' import 'selinux._selinux' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119ba6880> import 'selinux' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119596550> import ansible.module_utils.common.file # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/file.py import ansible.module_utils.common.process # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/process.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib/python3.9/site-packages/__pycache__/distro.cpython-39.pyc matches /usr/lib/python3.9/site-packages/distro.py # code object from '/usr/lib/python3.9/site-packages/__pycache__/distro.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/argparse.cpython-39.pyc matches /usr/lib64/python3.9/argparse.py # code object from '/usr/lib64/python3.9/__pycache__/argparse.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/gettext.cpython-39.pyc matches /usr/lib64/python3.9/gettext.py # code object from '/usr/lib64/python3.9/__pycache__/gettext.cpython-39.pyc' import 'gettext' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119bb7910> import 'argparse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111a143970> import 'distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119bea850> # destroy ansible.module_utils.distro import ansible.module_utils.distro # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/distro/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common._utils # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/_utils.py import ansible.module_utils.common.sys_info # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/common/sys_info.py import ansible.module_utils.basic # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/basic.py # zipimport: zlib available # zipimport: zlib available import ansible.modules # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/modules/__init__.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.namespace # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/namespace.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.compat.typing # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/compat/typing.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/multiprocessing/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/__init__.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/multiprocessing/__pycache__/context.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/context.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/context.cpython-39.pyc' # /usr/lib64/python3.9/multiprocessing/__pycache__/process.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/process.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/process.cpython-39.pyc' import 'multiprocessing.process' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111949bc70> # /usr/lib64/python3.9/multiprocessing/__pycache__/reduction.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/reduction.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/reduction.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/pickle.cpython-39.pyc matches /usr/lib64/python3.9/pickle.py # code object from '/usr/lib64/python3.9/__pycache__/pickle.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/_compat_pickle.cpython-39.pyc matches /usr/lib64/python3.9/_compat_pickle.py # code object from '/usr/lib64/python3.9/__pycache__/_compat_pickle.cpython-39.pyc' import '_compat_pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11196fda30> # extension module '_pickle' loaded from '/usr/lib64/python3.9/lib-dynload/_pickle.cpython-39-x86_64-linux-gnu.so' # extension module '_pickle' executed from '/usr/lib64/python3.9/lib-dynload/_pickle.cpython-39-x86_64-linux-gnu.so' import '_pickle' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f11196fd9a0> import 'pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119743b20> import 'multiprocessing.reduction' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119743550> import 'multiprocessing.context' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111972f2e0> import 'multiprocessing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111972f970> # /usr/lib64/python3.9/multiprocessing/__pycache__/pool.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/pool.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/pool.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/queue.cpython-39.pyc matches /usr/lib64/python3.9/queue.py # code object from '/usr/lib64/python3.9/__pycache__/queue.cpython-39.pyc' # extension module '_queue' loaded from '/usr/lib64/python3.9/lib-dynload/_queue.cpython-39-x86_64-linux-gnu.so' # extension module '_queue' executed from '/usr/lib64/python3.9/lib-dynload/_queue.cpython-39-x86_64-linux-gnu.so' import '_queue' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f11196e02b0> import 'queue' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11196e0a00> # /usr/lib64/python3.9/multiprocessing/__pycache__/util.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/util.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/util.cpython-39.pyc' import 'multiprocessing.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11196e0940> # /usr/lib64/python3.9/multiprocessing/__pycache__/connection.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/connection.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/connection.cpython-39.pyc' # extension module '_multiprocessing' loaded from '/usr/lib64/python3.9/lib-dynload/_multiprocessing.cpython-39-x86_64-linux-gnu.so' # extension module '_multiprocessing' executed from '/usr/lib64/python3.9/lib-dynload/_multiprocessing.cpython-39-x86_64-linux-gnu.so' import '_multiprocessing' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f11194fc0d0> import 'multiprocessing.connection' # <_frozen_importlib_external.SourceFileLoader object at 0x7f1119ba23a0> import 'multiprocessing.pool' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111972f670> import ansible.module_utils.facts.timeout # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/timeout.py import ansible.module_utils.facts.collector # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/collector.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.other # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/other/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.other.facter # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/other/facter.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.other.ohai # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/other/ohai.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.apparmor # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/apparmor.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.caps # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/caps.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.chroot # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/chroot.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.utils # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/utils.py import ansible.module_utils.facts.system.cmdline # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/cmdline.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.distribution # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/distribution.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.compat.datetime # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/compat/datetime.py import ansible.module_utils.facts.system.date_time # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/date_time.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.env # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/env.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.dns # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/dns.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.fips # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/fips.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.loadavg # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/loadavg.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/glob.cpython-39.pyc matches /usr/lib64/python3.9/glob.py # code object from '/usr/lib64/python3.9/__pycache__/glob.cpython-39.pyc' import 'glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11193eaeb0> # /usr/lib64/python3.9/__pycache__/configparser.cpython-39.pyc matches /usr/lib64/python3.9/configparser.py # code object from '/usr/lib64/python3.9/__pycache__/configparser.cpython-39.pyc' import 'configparser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11193ea9d0> import ansible.module_utils.facts.system.local # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/local.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.lsb # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/lsb.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.pkg_mgr # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/pkg_mgr.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.platform # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/platform.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/ssl.cpython-39.pyc matches /usr/lib64/python3.9/ssl.py # code object from '/usr/lib64/python3.9/__pycache__/ssl.cpython-39.pyc' # extension module '_ssl' loaded from '/usr/lib64/python3.9/lib-dynload/_ssl.cpython-39-x86_64-linux-gnu.so' # extension module '_ssl' executed from '/usr/lib64/python3.9/lib-dynload/_ssl.cpython-39-x86_64-linux-gnu.so' import '_ssl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f1119458bb0> import 'ssl' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11193fba60> import ansible.module_utils.facts.system.python # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/python.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.selinux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/selinux.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.compat.version # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/compat/version.py import ansible.module_utils.facts.system.service_mgr # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/service_mgr.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.ssh_pub_keys # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/ssh_pub_keys.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/getpass.cpython-39.pyc matches /usr/lib64/python3.9/getpass.py # code object from '/usr/lib64/python3.9/__pycache__/getpass.cpython-39.pyc' # extension module 'termios' loaded from '/usr/lib64/python3.9/lib-dynload/termios.cpython-39-x86_64-linux-gnu.so' # extension module 'termios' executed from '/usr/lib64/python3.9/lib-dynload/termios.cpython-39-x86_64-linux-gnu.so' import 'termios' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f111945e040> import 'getpass' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111945e6d0> import ansible.module_utils.facts.system.user # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/system/user.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.base # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/base.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.aix # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/aix.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.sysctl # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/sysctl.py import ansible.module_utils.facts.hardware.darwin # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/darwin.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.freebsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/freebsd.py import ansible.module_utils.facts.hardware.dragonfly # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/dragonfly.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.hpux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/hpux.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.linux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/linux.py import ansible.module_utils.facts.hardware.hurd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/hurd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.netbsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/netbsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.openbsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/openbsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.sunos # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/hardware/sunos.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.base # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/base.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.generic_bsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/generic_bsd.py import ansible.module_utils.facts.network.aix # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/aix.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.darwin # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/darwin.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.dragonfly # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/dragonfly.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.fc_wwn # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/fc_wwn.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.freebsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/freebsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.hpux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/hpux.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.hurd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/hurd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.linux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/linux.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.iscsi # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/iscsi.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.nvme # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/nvme.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.netbsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/netbsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.openbsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/openbsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.sunos # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/network/sunos.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.base # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/base.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.sysctl # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/sysctl.py import ansible.module_utils.facts.virtual.freebsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/freebsd.py import ansible.module_utils.facts.virtual.dragonfly # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/dragonfly.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.hpux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/hpux.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.linux # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/linux.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.netbsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/netbsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.openbsd # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/openbsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.sunos # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/virtual/sunos.py import ansible.module_utils.facts.default_collectors # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/default_collectors.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.ansible_collector # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/ansible_collector.py import ansible.module_utils.facts.compat # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/compat.py import ansible.module_utils.facts # loaded from Zip /tmp/ansible_ansible.legacy.setup_payload_1bl_1mta/ansible_ansible.legacy.setup_payload.zip/ansible/module_utils/facts/__init__.py # zipimport: zlib available import 'gc' # # /usr/lib64/python3.9/encodings/__pycache__/idna.cpython-39.pyc matches /usr/lib64/python3.9/encodings/idna.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/idna.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/stringprep.cpython-39.pyc matches /usr/lib64/python3.9/stringprep.py # code object from '/usr/lib64/python3.9/__pycache__/stringprep.cpython-39.pyc' # extension module 'unicodedata' loaded from '/usr/lib64/python3.9/lib-dynload/unicodedata.cpython-39-x86_64-linux-gnu.so' # extension module 'unicodedata' executed from '/usr/lib64/python3.9/lib-dynload/unicodedata.cpython-39-x86_64-linux-gnu.so' import 'unicodedata' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f11193e0190> import 'stringprep' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11193e01c0> import 'encodings.idna' # <_frozen_importlib_external.SourceFileLoader object at 0x7f111938ce80> # /usr/lib64/python3.9/multiprocessing/__pycache__/queues.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/queues.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/queues.cpython-39.pyc' import 'multiprocessing.queues' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11193e02b0> # /usr/lib64/python3.9/multiprocessing/__pycache__/synchronize.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/synchronize.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/synchronize.cpython-39.pyc' import 'multiprocessing.synchronize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11193a2730> # /usr/lib64/python3.9/multiprocessing/dummy/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/dummy/__init__.py # code object from '/usr/lib64/python3.9/multiprocessing/dummy/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/multiprocessing/dummy/__pycache__/connection.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/dummy/connection.py # code object from '/usr/lib64/python3.9/multiprocessing/dummy/__pycache__/connection.cpython-39.pyc' import 'multiprocessing.dummy.connection' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11191f2280> import 'multiprocessing.dummy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f11191f2070> PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame {"ansible_facts": {"ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "root", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_system": "Linux", "ansible_kernel": "5.14.0-511.el9.x86_64", "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Thu Sep 19 06:52:39 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.9.19", "ansible_fqdn": "managed-node2", "ansible_hostname": "managed-node2", "ansible_nodename": "managed-node2", "ansible_domain": "", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "e28ab0e542474a869c23f7ace4640799", "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_python": {"version": {"major": 3, "minor": 9, "micro": 19, "releaselevel": "final", "serial": 0}, "version_info": [3, 9, 19, "final", 0], "executable": "/usr/bin/python3.9", "has_sslcontext": true, "type": "cpython"}, "ansible_fibre_channel_wwn": [], "ansible_env": {"PYTHONVERBOSE": "1", "SHELL": "/bin/bash", "PWD": "/root", "LOGNAME": "root", "XDG_SESSION_TYPE": "tty", "_": "/usr/bin/python3.9", "MOTD_SHOWN": "pam", "HOME": "/root", "LANG": "en_US.UTF-8", "LS_COLORS": "", "SSH_CONNECTION": "10.31.14.85 48676 10.31.13.78 22", "XDG_SESSION_CLASS": "user", "SELINUX_ROLE_REQUESTED": "", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "USER": "root", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "1", "XDG_SESSION_ID": "5", "XDG_RUNTIME_DIR": "/run/user/0", "SSH_CLIENT": "10.31.14.85 48676 22", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "which_declare": "declare -f", "PATH": "/root/.local/bin:/root/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "SELINUX_LEVEL_REQUESTED": "", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "SSH_TTY": "/dev/pts/0", "BASH_FUNC_which%%": "() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz", "1", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz"], "ansible_processor_count": 1, "ansible_processor_cores": 1, "ansible_processor_threads_per_core": 2, "ansible_processor_vcpus": 2, "ansible_processor_nproc": 2, "ansible_memtotal_mb": 3532, "ansible_memfree_mb": 2785, "ansible_swaptotal_mb": 0, "ansible_swapfree_mb": 0, "ansible_memory_mb": {"real": {"total": 3532, "used": 747, "free": 2785}, "nocache": {"free": 3253, "used": 279}, "swap": {"total": 0, "free": 0, "used": 0, "cached": 0}}, "ansible_bios_date": "08/24/2006", "ansible_bios_vendor": "Xen", "ansible_bios_version": "4.11.amazon", "ansible_board_asset_tag": "NA", "ansible_board_name": "NA", "ansible_board_serial": "NA", "ansible_board_vendor": "NA", "ansible_board_version": "NA", "ansible_chassis_asset_tag": "NA", "ansible_chassis_serial": "NA", "ansible_chassis_vendor": "Xen", "ansible_chassis_version": "NA", "ansible_form_factor": "Other", "ansible_product_name": "HVM domU", "ansible_product_serial": "ec243623-fa66-7445-44ba-1070930583a9", "ansible_product_uuid": "ec243623-fa66-7445-44ba-1070930583a9", "ansible_product_version": "4.11.amazon", "ansible_system_vendor": "Xen", "ansible_devices": {"xvda": {"virtual": 1, "links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "vendor": null, "model": null, "sas_address": null, "sas_device_handle": null, "removable": "0", "support_discard": "512", "partitions": {"xvda1": {"links": {"ids": [], "uuids": ["ad406aa3-aab4-4a6a-aa73-3e870a6316ae"], "labels": [], "masters": []}, "start": "2048", "sectors": "524285919", "sectorsize": 512, "size": "250.00 GB", "uuid": "ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "holders": []}}, "rotational": "0", "scheduler_mode": "mq-deadline", "sectors": "524288000", "sectorsize": "512", "size": "250.00 GB", "host": "", "holders": []}}, "ansible_device_links": {"ids": {}, "uuids": {"xvda1": ["ad406aa3-aab4-4a6a-aa73-3e870a6316ae"]}, "labels": {}, "masters": {}}, "ansible_uptime_seconds": 426, "ansible_lvm": "N/A", "ansible_mounts": [{"mount": "/", "device": "/dev/xvda1", "fstype": "xfs", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "dump": 0, "passno": 0, "size_total": 268367278080, "size_available": 264273084416, "block_size": 4096, "block_total": 65519355, "block_available": 64519796, "block_used": 999559, "inode_total": 131071472, "inode_available": 130998253, "inode_used": 73219, "uuid": "ad406aa3-aab4-4a6a-aa73-3e870a6316ae"}], "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAPleAC0mV69PNpLSbmzZvoLD9LsCBzX6IHRLXV1uktk0r66T6Y57EoVgflJTdo6yU0zTaJjonNzFmvC69tiRsCyywGjnvnBOvIH2vrgNGCUdVYPZbbtmQlJvol7NFFfyXQR4RSPqBKT67rYbCzbETM4j+bdDgTeDk6l7wXwz9RVvAAAAFQCuAyyjbOBDKyIW26LGcI9/nmWpHwAAAIEApIE1W6KQ7qs5kJXBdSaPoWaZUxuQhXkPWORFe7/MBn5SojDfxvJjFPo6t4QsovaCnm532Zghh1ZdB0pNm0vYcRbz3wMdfMucw/KHWt6ZEtI+sLwuMyhAVEXzmE34iXkyePtELiYzY6NyxuJ04IujI9UwD7ZnqFBHVFz529oXikIAAACBAPdUu+4Qo82CMcmrGD9vNUgtsts6GCjqBDuov8GJEALZ9ZNLlyVoNtBHLMQH9e0czLygyNGw/IDosRQkKdX4Vh4A7KXujTIOyytaN4JVJCuOBY/PeX4lreAO/UTTUJ27yT/J0Oy2Hbt+d8fZnTkZReRNPFCzvdb1nuPMG5nAyQtL", "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQCzkKXWiNuOrU77QQcZuT2T9XVh655Sh8Sv9vLWLa1uj7ceaNsB0TBiqvDFvYPENhdKceYaGAFU7sjqbmp5dlivYwPBiBWvcOgqnpBqrMG5SvP1RMiORpW6GupBLnUaMVjopPLIi0/CDlSl2eODcEnQI6BpxCCSedEKU9UrRrCFJy+6KPQXepPwKwPTd1TMzO8wpo57B5MYrjnquTNxMfgBkYsHB/V77d0tKq8qGBTkAPD8wEWLIcZOI+SyYEfCraQ95dOGAPRTFijnd7S15CugSlJ/vvcHSFXOlbgFzeNnU2jZneagkBfaOJch72opD3ebISSHCx1/kJvHN7MbksI+ljJa3Nw5LwP1XjUpT7dQMOZJDdVStXKp86K4XpWud+wMbQVVyU5QoFsCl7YTWWmSDRiPJOQI2myfizCT8i42rJ0WXm5OnqpHn1Jw4nGlcVnfgPQA/zxMldzReXdHnvriqKC9+97XgY6pj42YYP78PhOu1D2xH1AXmloNM+63VvU=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPT1h7wNcUomxtav688iXvnCnFqrHnEKf4gRaBY3w4BwbWOGxE8hq5snF9Tp+0agFeN/u980/y8BJWdWIO9Lz8I=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIPe8liWy3mh5GzCz9W616J2ArXnLOjLOZSwfmBX3Q1SI", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "9", "ansible_distribution_major_version": "9", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_is_chroot": false, "ansible_apparmor": {"status": "disabled"}, "ansible_iscsi_iqn": "", "ansible_virtualization_type": "xen", "ansible_virtualization_role": "guest", "ansible_virtualization_tech_guest": ["xen"], "ansible_virtualization_tech_host": [], "ansible_hostnqn": "nqn.2014-08.org.nvmexpress:uuid:d5aef1ea-3141-48ae-bf33-0c6b351dd422", "ansible_fips": false, "ansible_pkg_mgr": "dnf", "ansible_service_mgr": "systemd", "ansible_loadavg": {"1m": 0.38, "5m": 0.26, "15m": 0.13}, "ansible_interfaces": ["lo", "eth0"], "ansible_eth0": {"device": "eth0", "macaddress": "0a:ff:ff:f5:f2:b9", "mtu": 9001, "active": true, "module": "xen_netfront", "type": "ether", "pciid": "vif-0", "promisc": false, "ipv4": {"address": "10.31.13.78", "broadcast": "10.31.15.255", "netmask": "255.255.252.0", "network": "10.31.12.0", "prefix": "22"}, "ipv6": [{"address": "fe80::8ff:ffff:fef5:f2b9", "prefix": "64", "scope": "link"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "on [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv6": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on", "tx_scatter_gather_fraglist": "off [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "off [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "off [fixed]", "tx_lockless": "off [fixed]", "netns_local": "off [fixed]", "tx_gso_robust": "on [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_gso_list": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off", "loopback": "off [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_lo": {"device": "lo", "mtu": 65536, "active": true, "type": "loopback", "promisc": false, "ipv4": {"address": "127.0.0.1", "broadcast": "", "netmask": "255.0.0.0", "network": "127.0.0.0", "prefix": "8"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "on [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "on [fixed]", "tx_lockless": "on [fixed]", "netns_local": "on [fixed]", "tx_gso_robust": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "on", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "on", "tx_gso_list": "on", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off [fixed]", "loopback": "on [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_default_ipv4": {"gateway": "10.31.12.1", "interface": "eth0", "address": "10.31.13.78", "broadcast": "10.31.15.255", "netmask": "255.255.252.0", "network": "10.31.12.0", "prefix": "22", "macaddress": "0a:ff:ff:f5:f2:b9", "mtu": 9001, "type": "ether", "alias": "eth0"}, "ansible_default_ipv6": {}, "ansible_all_ipv4_addresses": ["10.31.13.78"], "ansible_all_ipv6_addresses": ["fe80::8ff:ffff:fef5:f2b9"], "ansible_locally_reachable_ips": {"ipv4": ["10.31.13.78", "127.0.0.0/8", "127.0.0.1"], "ipv6": ["::1", "fe80::8ff:ffff:fef5:f2b9"]}, "ansible_lsb": {}, "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-5.14.0-511.el9.x86_64", "root": "UUID=ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": "ttyS0,115200n8"}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-5.14.0-511.el9.x86_64", "root": "UUID=ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": ["tty0", "ttyS0,115200n8"]}, "ansible_local": {}, "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Tuesday", "weekday_number": "2", "weeknumber": "39", "day": "24", "hour": "14", "minute": "54", "second": "24", "epoch": "1727204064", "epoch_int": "1727204064", "date": "2024-09-24", "time": "14:54:24", "iso8601_micro": "2024-09-24T18:54:24.007740Z", "iso8601": "2024-09-24T18:54:24Z", "iso8601_basic": "20240924T145424007740", "iso8601_basic_short": "20240924T145424", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "gather_subset": ["all"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["all"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.path_hooks # clear sys.path_importer_cache # clear sys.meta_path # clear sys.__interactivehook__ # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing encodings.latin_1 # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing _locale # cleanup[2] removing _bootlocale # destroy _bootlocale # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing sre_constants # destroy sre_constants # cleanup[2] removing sre_parse # cleanup[2] removing sre_compile # cleanup[2] removing _heapq # cleanup[2] removing heapq # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing collections.abc # cleanup[2] removing contextlib # cleanup[2] removing typing # destroy typing # cleanup[2] removing importlib.abc # cleanup[2] removing importlib.util # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing threading # cleanup[2] removing _bz2 # destroy _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing pwd # cleanup[2] removing grp # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing tempfile # cleanup[2] removing zipfile # destroy zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing fcntl # cleanup[2] removing locale # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing subprocess # cleanup[2] removing platform # cleanup[2] removing shlex # cleanup[2] removing token # destroy token # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # cleanup[2] removing ansible.module_utils.facts.namespace # cleanup[2] removing ansible.module_utils.compat.typing # cleanup[2] removing multiprocessing.process # cleanup[2] removing _compat_pickle # cleanup[2] removing _pickle # cleanup[2] removing pickle # cleanup[2] removing multiprocessing.reduction # cleanup[2] removing multiprocessing.context # cleanup[2] removing __mp_main__ # destroy __main__ # cleanup[2] removing multiprocessing # cleanup[2] removing _queue # cleanup[2] removing queue # cleanup[2] removing multiprocessing.util # cleanup[2] removing _multiprocessing # cleanup[2] removing multiprocessing.connection # cleanup[2] removing multiprocessing.pool # cleanup[2] removing ansible.module_utils.facts.timeout # cleanup[2] removing ansible.module_utils.facts.collector # cleanup[2] removing ansible.module_utils.facts.other # cleanup[2] removing ansible.module_utils.facts.other.facter # cleanup[2] removing ansible.module_utils.facts.other.ohai # cleanup[2] removing ansible.module_utils.facts.system # cleanup[2] removing ansible.module_utils.facts.system.apparmor # cleanup[2] removing ansible.module_utils.facts.system.caps # cleanup[2] removing ansible.module_utils.facts.system.chroot # cleanup[2] removing ansible.module_utils.facts.utils # cleanup[2] removing ansible.module_utils.facts.system.cmdline # cleanup[2] removing ansible.module_utils.facts.system.distribution # cleanup[2] removing ansible.module_utils.compat.datetime # destroy ansible.module_utils.compat.datetime # cleanup[2] removing ansible.module_utils.facts.system.date_time # cleanup[2] removing ansible.module_utils.facts.system.env # cleanup[2] removing ansible.module_utils.facts.system.dns # cleanup[2] removing ansible.module_utils.facts.system.fips # cleanup[2] removing ansible.module_utils.facts.system.loadavg # cleanup[2] removing glob # cleanup[2] removing configparser # cleanup[2] removing ansible.module_utils.facts.system.local # cleanup[2] removing ansible.module_utils.facts.system.lsb # cleanup[2] removing ansible.module_utils.facts.system.pkg_mgr # cleanup[2] removing ansible.module_utils.facts.system.platform # cleanup[2] removing _ssl # cleanup[2] removing ssl # destroy ssl # cleanup[2] removing ansible.module_utils.facts.system.python # cleanup[2] removing ansible.module_utils.facts.system.selinux # cleanup[2] removing ansible.module_utils.compat.version # destroy ansible.module_utils.compat.version # cleanup[2] removing ansible.module_utils.facts.system.service_mgr # cleanup[2] removing ansible.module_utils.facts.system.ssh_pub_keys # cleanup[2] removing termios # cleanup[2] removing getpass # cleanup[2] removing ansible.module_utils.facts.system.user # cleanup[2] removing ansible.module_utils.facts.hardware # cleanup[2] removing ansible.module_utils.facts.hardware.base # cleanup[2] removing ansible.module_utils.facts.hardware.aix # cleanup[2] removing ansible.module_utils.facts.sysctl # cleanup[2] removing ansible.module_utils.facts.hardware.darwin # cleanup[2] removing ansible.module_utils.facts.hardware.freebsd # cleanup[2] removing ansible.module_utils.facts.hardware.dragonfly # cleanup[2] removing ansible.module_utils.facts.hardware.hpux # cleanup[2] removing ansible.module_utils.facts.hardware.linux # cleanup[2] removing ansible.module_utils.facts.hardware.hurd # cleanup[2] removing ansible.module_utils.facts.hardware.netbsd # cleanup[2] removing ansible.module_utils.facts.hardware.openbsd # cleanup[2] removing ansible.module_utils.facts.hardware.sunos # cleanup[2] removing ansible.module_utils.facts.network # cleanup[2] removing ansible.module_utils.facts.network.base # cleanup[2] removing ansible.module_utils.facts.network.generic_bsd # cleanup[2] removing ansible.module_utils.facts.network.aix # cleanup[2] removing ansible.module_utils.facts.network.darwin # cleanup[2] removing ansible.module_utils.facts.network.dragonfly # cleanup[2] removing ansible.module_utils.facts.network.fc_wwn # cleanup[2] removing ansible.module_utils.facts.network.freebsd # cleanup[2] removing ansible.module_utils.facts.network.hpux # cleanup[2] removing ansible.module_utils.facts.network.hurd # cleanup[2] removing ansible.module_utils.facts.network.linux # cleanup[2] removing ansible.module_utils.facts.network.iscsi # cleanup[2] removing ansible.module_utils.facts.network.nvme # cleanup[2] removing ansible.module_utils.facts.network.netbsd # cleanup[2] removing ansible.module_utils.facts.network.openbsd # cleanup[2] removing ansible.module_utils.facts.network.sunos # cleanup[2] removing ansible.module_utils.facts.virtual # cleanup[2] removing ansible.module_utils.facts.virtual.base # cleanup[2] removing ansible.module_utils.facts.virtual.sysctl # cleanup[2] removing ansible.module_utils.facts.virtual.freebsd # cleanup[2] removing ansible.module_utils.facts.virtual.dragonfly # cleanup[2] removing ansible.module_utils.facts.virtual.hpux # cleanup[2] removing ansible.module_utils.facts.virtual.linux # cleanup[2] removing ansible.module_utils.facts.virtual.netbsd # cleanup[2] removing ansible.module_utils.facts.virtual.openbsd # cleanup[2] removing ansible.module_utils.facts.virtual.sunos # cleanup[2] removing ansible.module_utils.facts.default_collectors # cleanup[2] removing ansible.module_utils.facts.ansible_collector # cleanup[2] removing ansible.module_utils.facts.compat # cleanup[2] removing ansible.module_utils.facts # destroy ansible.module_utils.facts # destroy ansible.module_utils.facts.namespace # destroy ansible.module_utils.facts.other # destroy ansible.module_utils.facts.other.facter # destroy ansible.module_utils.facts.other.ohai # destroy ansible.module_utils.facts.system # destroy ansible.module_utils.facts.system.apparmor # destroy ansible.module_utils.facts.system.caps # destroy ansible.module_utils.facts.system.chroot # destroy ansible.module_utils.facts.system.cmdline # destroy ansible.module_utils.facts.system.distribution # destroy ansible.module_utils.facts.system.date_time # destroy ansible.module_utils.facts.system.env # destroy ansible.module_utils.facts.system.dns # destroy ansible.module_utils.facts.system.fips # destroy ansible.module_utils.facts.system.loadavg # destroy ansible.module_utils.facts.system.local # destroy ansible.module_utils.facts.system.lsb # destroy ansible.module_utils.facts.system.pkg_mgr # destroy ansible.module_utils.facts.system.platform # destroy ansible.module_utils.facts.system.python # destroy ansible.module_utils.facts.system.selinux # destroy ansible.module_utils.facts.system.service_mgr # destroy ansible.module_utils.facts.system.ssh_pub_keys # destroy ansible.module_utils.facts.system.user # destroy ansible.module_utils.facts.utils # destroy ansible.module_utils.facts.hardware # destroy ansible.module_utils.facts.hardware.base # destroy ansible.module_utils.facts.hardware.aix # destroy ansible.module_utils.facts.hardware.darwin # destroy ansible.module_utils.facts.hardware.freebsd # destroy ansible.module_utils.facts.hardware.dragonfly # destroy ansible.module_utils.facts.hardware.hpux # destroy ansible.module_utils.facts.hardware.linux # destroy ansible.module_utils.facts.hardware.hurd # destroy ansible.module_utils.facts.hardware.netbsd # destroy ansible.module_utils.facts.hardware.openbsd # destroy ansible.module_utils.facts.hardware.sunos # destroy ansible.module_utils.facts.sysctl # destroy ansible.module_utils.facts.network # destroy ansible.module_utils.facts.network.base # destroy ansible.module_utils.facts.network.generic_bsd # destroy ansible.module_utils.facts.network.aix # destroy ansible.module_utils.facts.network.darwin # destroy ansible.module_utils.facts.network.dragonfly # destroy ansible.module_utils.facts.network.fc_wwn # destroy ansible.module_utils.facts.network.freebsd # destroy ansible.module_utils.facts.network.hpux # destroy ansible.module_utils.facts.network.hurd # destroy ansible.module_utils.facts.network.linux # destroy ansible.module_utils.facts.network.iscsi # destroy ansible.module_utils.facts.network.nvme # destroy ansible.module_utils.facts.network.netbsd # destroy ansible.module_utils.facts.network.openbsd # destroy ansible.module_utils.facts.network.sunos # destroy ansible.module_utils.facts.virtual # destroy ansible.module_utils.facts.virtual.base # destroy ansible.module_utils.facts.virtual.sysctl # destroy ansible.module_utils.facts.virtual.freebsd # destroy ansible.module_utils.facts.virtual.dragonfly # destroy ansible.module_utils.facts.virtual.hpux # destroy ansible.module_utils.facts.virtual.linux # destroy ansible.module_utils.facts.virtual.netbsd # destroy ansible.module_utils.facts.virtual.openbsd # destroy ansible.module_utils.facts.virtual.sunos # destroy ansible.module_utils.facts.compat # cleanup[2] removing gc # cleanup[2] removing unicodedata # cleanup[2] removing stringprep # cleanup[2] removing encodings.idna # cleanup[2] removing multiprocessing.queues # cleanup[2] removing multiprocessing.synchronize # cleanup[2] removing multiprocessing.dummy.connection # cleanup[2] removing multiprocessing.dummy # destroy _sitebuiltins # destroy importlib.util # destroy importlib.abc # destroy importlib.machinery # destroy zipimport # destroy _compression # destroy binascii # destroy importlib # destroy bz2 # destroy lzma # destroy __main__ # destroy locale # destroy systemd.journal # destroy systemd.daemon # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy encodings # destroy syslog # destroy uuid # destroy selinux # destroy distro # destroy logging # destroy argparse # destroy ansible.module_utils.facts.default_collectors # destroy ansible.module_utils.facts.ansible_collector # destroy multiprocessing # destroy multiprocessing.queues # destroy multiprocessing.synchronize # destroy multiprocessing.dummy # destroy multiprocessing.pool # destroy pickle # destroy _compat_pickle # destroy queue # destroy multiprocessing.reduction # destroy shlex # destroy datetime # destroy base64 # destroy ansible.module_utils.compat.selinux # destroy getpass # destroy json # destroy socket # destroy struct # destroy glob # destroy ansible.module_utils.compat.typing # destroy ansible.module_utils.facts.timeout # destroy ansible.module_utils.facts.collector # destroy multiprocessing.connection # destroy tempfile # destroy multiprocessing.context # destroy multiprocessing.process # destroy multiprocessing.util # destroy array # destroy multiprocessing.dummy.connection # cleanup[3] wiping encodings.idna # destroy stringprep # cleanup[3] wiping unicodedata # cleanup[3] wiping gc # cleanup[3] wiping termios # cleanup[3] wiping _ssl # cleanup[3] wiping configparser # cleanup[3] wiping _multiprocessing # cleanup[3] wiping _queue # cleanup[3] wiping _pickle # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # destroy configparser # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # cleanup[3] wiping tokenize # cleanup[3] wiping platform # destroy subprocess # cleanup[3] wiping selectors # cleanup[3] wiping select # cleanup[3] wiping _posixsubprocess # cleanup[3] wiping signal # cleanup[3] wiping fcntl # cleanup[3] wiping atexit # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping _blake2 # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping shutil # destroy fnmatch # cleanup[3] wiping grp # cleanup[3] wiping pwd # cleanup[3] wiping _lzma # cleanup[3] wiping threading # cleanup[3] wiping zlib # cleanup[3] wiping errno # cleanup[3] wiping weakref # cleanup[3] wiping contextlib # cleanup[3] wiping collections.abc # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy enum # destroy sre_compile # destroy copyreg # cleanup[3] wiping functools # cleanup[3] wiping _functools # destroy _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy heapq # destroy collections.abc # cleanup[3] wiping _collections # destroy _collections # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping itertools # cleanup[3] wiping _heapq # cleanup[3] wiping sre_parse # cleanup[3] wiping _sre # cleanup[3] wiping types # cleanup[3] wiping _locale # destroy _locale # cleanup[3] wiping os # cleanup[3] wiping os.path # destroy genericpath # cleanup[3] wiping posixpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.latin_1 # cleanup[3] wiping _signal # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy unicodedata # destroy gc # destroy termios # destroy _ssl # destroy _multiprocessing # destroy _queue # destroy _pickle # destroy systemd._daemon # destroy _socket # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy fcntl # destroy _blake2 # destroy _lzma # destroy zlib # destroy _signal # destroy platform # destroy _uuid # destroy _sre # destroy sre_parse # destroy tokenize # destroy _heapq # destroy posixpath # destroy stat # destroy ansible.module_utils.six.moves.urllib # destroy errno # destroy signal # destroy contextlib # destroy pwd # destroy grp # destroy _posixsubprocess # destroy selectors # destroy select # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy itertools # destroy operator # destroy ansible.module_utils.six.moves # destroy _operator # destroy _frozen_importlib_external # destroy _imp # destroy io # destroy marshal # destroy _frozen_importlib # clear sys.audit hooks , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. [WARNING]: Module invocation had junk after the JSON data: # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.path_hooks # clear sys.path_importer_cache # clear sys.meta_path # clear sys.__interactivehook__ # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing encodings.latin_1 # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing _locale # cleanup[2] removing _bootlocale # destroy _bootlocale # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing sre_constants # destroy sre_constants # cleanup[2] removing sre_parse # cleanup[2] removing sre_compile # cleanup[2] removing _heapq # cleanup[2] removing heapq # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing collections.abc # cleanup[2] removing contextlib # cleanup[2] removing typing # destroy typing # cleanup[2] removing importlib.abc # cleanup[2] removing importlib.util # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing threading # cleanup[2] removing _bz2 # destroy _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing pwd # cleanup[2] removing grp # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing tempfile # cleanup[2] removing zipfile # destroy zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing fcntl # cleanup[2] removing locale # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing subprocess # cleanup[2] removing platform # cleanup[2] removing shlex # cleanup[2] removing token # destroy token # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # cleanup[2] removing ansible.module_utils.facts.namespace # cleanup[2] removing ansible.module_utils.compat.typing # cleanup[2] removing multiprocessing.process # cleanup[2] removing _compat_pickle # cleanup[2] removing _pickle # cleanup[2] removing pickle # cleanup[2] removing multiprocessing.reduction # cleanup[2] removing multiprocessing.context # cleanup[2] removing __mp_main__ # destroy __main__ # cleanup[2] removing multiprocessing # cleanup[2] removing _queue # cleanup[2] removing queue # cleanup[2] removing multiprocessing.util # cleanup[2] removing _multiprocessing # cleanup[2] removing multiprocessing.connection # cleanup[2] removing multiprocessing.pool # cleanup[2] removing ansible.module_utils.facts.timeout # cleanup[2] removing ansible.module_utils.facts.collector # cleanup[2] removing ansible.module_utils.facts.other # cleanup[2] removing ansible.module_utils.facts.other.facter # cleanup[2] removing ansible.module_utils.facts.other.ohai # cleanup[2] removing ansible.module_utils.facts.system # cleanup[2] removing ansible.module_utils.facts.system.apparmor # cleanup[2] removing ansible.module_utils.facts.system.caps # cleanup[2] removing ansible.module_utils.facts.system.chroot # cleanup[2] removing ansible.module_utils.facts.utils # cleanup[2] removing ansible.module_utils.facts.system.cmdline # cleanup[2] removing ansible.module_utils.facts.system.distribution # cleanup[2] removing ansible.module_utils.compat.datetime # destroy ansible.module_utils.compat.datetime # cleanup[2] removing ansible.module_utils.facts.system.date_time # cleanup[2] removing ansible.module_utils.facts.system.env # cleanup[2] removing ansible.module_utils.facts.system.dns # cleanup[2] removing ansible.module_utils.facts.system.fips # cleanup[2] removing ansible.module_utils.facts.system.loadavg # cleanup[2] removing glob # cleanup[2] removing configparser # cleanup[2] removing ansible.module_utils.facts.system.local # cleanup[2] removing ansible.module_utils.facts.system.lsb # cleanup[2] removing ansible.module_utils.facts.system.pkg_mgr # cleanup[2] removing ansible.module_utils.facts.system.platform # cleanup[2] removing _ssl # cleanup[2] removing ssl # destroy ssl # cleanup[2] removing ansible.module_utils.facts.system.python # cleanup[2] removing ansible.module_utils.facts.system.selinux # cleanup[2] removing ansible.module_utils.compat.version # destroy ansible.module_utils.compat.version # cleanup[2] removing ansible.module_utils.facts.system.service_mgr # cleanup[2] removing ansible.module_utils.facts.system.ssh_pub_keys # cleanup[2] removing termios # cleanup[2] removing getpass # cleanup[2] removing ansible.module_utils.facts.system.user # cleanup[2] removing ansible.module_utils.facts.hardware # cleanup[2] removing ansible.module_utils.facts.hardware.base # cleanup[2] removing ansible.module_utils.facts.hardware.aix # cleanup[2] removing ansible.module_utils.facts.sysctl # cleanup[2] removing ansible.module_utils.facts.hardware.darwin # cleanup[2] removing ansible.module_utils.facts.hardware.freebsd # cleanup[2] removing ansible.module_utils.facts.hardware.dragonfly # cleanup[2] removing ansible.module_utils.facts.hardware.hpux # cleanup[2] removing ansible.module_utils.facts.hardware.linux # cleanup[2] removing ansible.module_utils.facts.hardware.hurd # cleanup[2] removing ansible.module_utils.facts.hardware.netbsd # cleanup[2] removing ansible.module_utils.facts.hardware.openbsd # cleanup[2] removing ansible.module_utils.facts.hardware.sunos # cleanup[2] removing ansible.module_utils.facts.network # cleanup[2] removing ansible.module_utils.facts.network.base # cleanup[2] removing ansible.module_utils.facts.network.generic_bsd # cleanup[2] removing ansible.module_utils.facts.network.aix # cleanup[2] removing ansible.module_utils.facts.network.darwin # cleanup[2] removing ansible.module_utils.facts.network.dragonfly # cleanup[2] removing ansible.module_utils.facts.network.fc_wwn # cleanup[2] removing ansible.module_utils.facts.network.freebsd # cleanup[2] removing ansible.module_utils.facts.network.hpux # cleanup[2] removing ansible.module_utils.facts.network.hurd # cleanup[2] removing ansible.module_utils.facts.network.linux # cleanup[2] removing ansible.module_utils.facts.network.iscsi # cleanup[2] removing ansible.module_utils.facts.network.nvme # cleanup[2] removing ansible.module_utils.facts.network.netbsd # cleanup[2] removing ansible.module_utils.facts.network.openbsd # cleanup[2] removing ansible.module_utils.facts.network.sunos # cleanup[2] removing ansible.module_utils.facts.virtual # cleanup[2] removing ansible.module_utils.facts.virtual.base # cleanup[2] removing ansible.module_utils.facts.virtual.sysctl # cleanup[2] removing ansible.module_utils.facts.virtual.freebsd # cleanup[2] removing ansible.module_utils.facts.virtual.dragonfly # cleanup[2] removing ansible.module_utils.facts.virtual.hpux # cleanup[2] removing ansible.module_utils.facts.virtual.linux # cleanup[2] removing ansible.module_utils.facts.virtual.netbsd # cleanup[2] removing ansible.module_utils.facts.virtual.openbsd # cleanup[2] removing ansible.module_utils.facts.virtual.sunos # cleanup[2] removing ansible.module_utils.facts.default_collectors # cleanup[2] removing ansible.module_utils.facts.ansible_collector # cleanup[2] removing ansible.module_utils.facts.compat # cleanup[2] removing ansible.module_utils.facts # destroy ansible.module_utils.facts # destroy ansible.module_utils.facts.namespace # destroy ansible.module_utils.facts.other # destroy ansible.module_utils.facts.other.facter # destroy ansible.module_utils.facts.other.ohai # destroy ansible.module_utils.facts.system # destroy ansible.module_utils.facts.system.apparmor # destroy ansible.module_utils.facts.system.caps # destroy ansible.module_utils.facts.system.chroot # destroy ansible.module_utils.facts.system.cmdline # destroy ansible.module_utils.facts.system.distribution # destroy ansible.module_utils.facts.system.date_time # destroy ansible.module_utils.facts.system.env # destroy ansible.module_utils.facts.system.dns # destroy ansible.module_utils.facts.system.fips # destroy ansible.module_utils.facts.system.loadavg # destroy ansible.module_utils.facts.system.local # destroy ansible.module_utils.facts.system.lsb # destroy ansible.module_utils.facts.system.pkg_mgr # destroy ansible.module_utils.facts.system.platform # destroy ansible.module_utils.facts.system.python # destroy ansible.module_utils.facts.system.selinux # destroy ansible.module_utils.facts.system.service_mgr # destroy ansible.module_utils.facts.system.ssh_pub_keys # destroy ansible.module_utils.facts.system.user # destroy ansible.module_utils.facts.utils # destroy ansible.module_utils.facts.hardware # destroy ansible.module_utils.facts.hardware.base # destroy ansible.module_utils.facts.hardware.aix # destroy ansible.module_utils.facts.hardware.darwin # destroy ansible.module_utils.facts.hardware.freebsd # destroy ansible.module_utils.facts.hardware.dragonfly # destroy ansible.module_utils.facts.hardware.hpux # destroy ansible.module_utils.facts.hardware.linux # destroy ansible.module_utils.facts.hardware.hurd # destroy ansible.module_utils.facts.hardware.netbsd # destroy ansible.module_utils.facts.hardware.openbsd # destroy ansible.module_utils.facts.hardware.sunos # destroy ansible.module_utils.facts.sysctl # destroy ansible.module_utils.facts.network # destroy ansible.module_utils.facts.network.base # destroy ansible.module_utils.facts.network.generic_bsd # destroy ansible.module_utils.facts.network.aix # destroy ansible.module_utils.facts.network.darwin # destroy ansible.module_utils.facts.network.dragonfly # destroy ansible.module_utils.facts.network.fc_wwn # destroy ansible.module_utils.facts.network.freebsd # destroy ansible.module_utils.facts.network.hpux # destroy ansible.module_utils.facts.network.hurd # destroy ansible.module_utils.facts.network.linux # destroy ansible.module_utils.facts.network.iscsi # destroy ansible.module_utils.facts.network.nvme # destroy ansible.module_utils.facts.network.netbsd # destroy ansible.module_utils.facts.network.openbsd # destroy ansible.module_utils.facts.network.sunos # destroy ansible.module_utils.facts.virtual # destroy ansible.module_utils.facts.virtual.base # destroy ansible.module_utils.facts.virtual.sysctl # destroy ansible.module_utils.facts.virtual.freebsd # destroy ansible.module_utils.facts.virtual.dragonfly # destroy ansible.module_utils.facts.virtual.hpux # destroy ansible.module_utils.facts.virtual.linux # destroy ansible.module_utils.facts.virtual.netbsd # destroy ansible.module_utils.facts.virtual.openbsd # destroy ansible.module_utils.facts.virtual.sunos # destroy ansible.module_utils.facts.compat # cleanup[2] removing gc # cleanup[2] removing unicodedata # cleanup[2] removing stringprep # cleanup[2] removing encodings.idna # cleanup[2] removing multiprocessing.queues # cleanup[2] removing multiprocessing.synchronize # cleanup[2] removing multiprocessing.dummy.connection # cleanup[2] removing multiprocessing.dummy # destroy _sitebuiltins # destroy importlib.util # destroy importlib.abc # destroy importlib.machinery # destroy zipimport # destroy _compression # destroy binascii # destroy importlib # destroy bz2 # destroy lzma # destroy __main__ # destroy locale # destroy systemd.journal # destroy systemd.daemon # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy encodings # destroy syslog # destroy uuid # destroy selinux # destroy distro # destroy logging # destroy argparse # destroy ansible.module_utils.facts.default_collectors # destroy ansible.module_utils.facts.ansible_collector # destroy multiprocessing # destroy multiprocessing.queues # destroy multiprocessing.synchronize # destroy multiprocessing.dummy # destroy multiprocessing.pool # destroy pickle # destroy _compat_pickle # destroy queue # destroy multiprocessing.reduction # destroy shlex # destroy datetime # destroy base64 # destroy ansible.module_utils.compat.selinux # destroy getpass # destroy json # destroy socket # destroy struct # destroy glob # destroy ansible.module_utils.compat.typing # destroy ansible.module_utils.facts.timeout # destroy ansible.module_utils.facts.collector # destroy multiprocessing.connection # destroy tempfile # destroy multiprocessing.context # destroy multiprocessing.process # destroy multiprocessing.util # destroy array # destroy multiprocessing.dummy.connection # cleanup[3] wiping encodings.idna # destroy stringprep # cleanup[3] wiping unicodedata # cleanup[3] wiping gc # cleanup[3] wiping termios # cleanup[3] wiping _ssl # cleanup[3] wiping configparser # cleanup[3] wiping _multiprocessing # cleanup[3] wiping _queue # cleanup[3] wiping _pickle # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # destroy configparser # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # cleanup[3] wiping tokenize # cleanup[3] wiping platform # destroy subprocess # cleanup[3] wiping selectors # cleanup[3] wiping select # cleanup[3] wiping _posixsubprocess # cleanup[3] wiping signal # cleanup[3] wiping fcntl # cleanup[3] wiping atexit # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping _blake2 # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping shutil # destroy fnmatch # cleanup[3] wiping grp # cleanup[3] wiping pwd # cleanup[3] wiping _lzma # cleanup[3] wiping threading # cleanup[3] wiping zlib # cleanup[3] wiping errno # cleanup[3] wiping weakref # cleanup[3] wiping contextlib # cleanup[3] wiping collections.abc # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy enum # destroy sre_compile # destroy copyreg # cleanup[3] wiping functools # cleanup[3] wiping _functools # destroy _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy heapq # destroy collections.abc # cleanup[3] wiping _collections # destroy _collections # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping itertools # cleanup[3] wiping _heapq # cleanup[3] wiping sre_parse # cleanup[3] wiping _sre # cleanup[3] wiping types # cleanup[3] wiping _locale # destroy _locale # cleanup[3] wiping os # cleanup[3] wiping os.path # destroy genericpath # cleanup[3] wiping posixpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.latin_1 # cleanup[3] wiping _signal # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy unicodedata # destroy gc # destroy termios # destroy _ssl # destroy _multiprocessing # destroy _queue # destroy _pickle # destroy systemd._daemon # destroy _socket # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy fcntl # destroy _blake2 # destroy _lzma # destroy zlib # destroy _signal # destroy platform # destroy _uuid # destroy _sre # destroy sre_parse # destroy tokenize # destroy _heapq # destroy posixpath # destroy stat # destroy ansible.module_utils.six.moves.urllib # destroy errno # destroy signal # destroy contextlib # destroy pwd # destroy grp # destroy _posixsubprocess # destroy selectors # destroy select # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy itertools # destroy operator # destroy ansible.module_utils.six.moves # destroy _operator # destroy _frozen_importlib_external # destroy _imp # destroy io # destroy marshal # destroy _frozen_importlib # clear sys.audit hooks [WARNING]: Platform linux on host managed-node2 is using the discovered Python interpreter at /usr/bin/python3.9, but future installation of another Python interpreter could change the meaning of that path. See https://docs.ansible.com/ansible- core/2.17/reference_appendices/interpreter_discovery.html for more information. 13118 1727204064.06270: done with _execute_module (ansible.legacy.setup, {'_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.setup', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204062.8121483-13211-79342327744476/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204064.06278: _low_level_execute_command(): starting 13118 1727204064.06285: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204062.8121483-13211-79342327744476/ > /dev/null 2>&1 && sleep 0' 13118 1727204064.08394: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204064.08414: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204064.08431: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204064.08452: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204064.08507: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204064.08521: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204064.08539: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204064.08556: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204064.08574: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204064.08586: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204064.08600: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204064.08614: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204064.08632: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204064.08645: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204064.08656: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204064.08670: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204064.08753: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204064.08917: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204064.08938: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204064.09016: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204064.10971: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204064.10975: stdout chunk (state=3): >>><<< 13118 1727204064.10978: stderr chunk (state=3): >>><<< 13118 1727204064.11073: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204064.11077: handler run complete 13118 1727204064.11272: variable 'ansible_facts' from source: unknown 13118 1727204064.11275: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204064.11604: variable 'ansible_facts' from source: unknown 13118 1727204064.11843: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204064.12095: attempt loop complete, returning result 13118 1727204064.12104: _execute() done 13118 1727204064.12110: dumping result to json 13118 1727204064.12262: done dumping result, returning 13118 1727204064.12278: done running TaskExecutor() for managed-node2/TASK: Gathering Facts [0affcd87-79f5-56a3-0a64-0000000000cc] 13118 1727204064.12288: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000cc ok: [managed-node2] 13118 1727204064.13755: no more pending results, returning what we have 13118 1727204064.13759: results queue empty 13118 1727204064.13761: checking for any_errors_fatal 13118 1727204064.13762: done checking for any_errors_fatal 13118 1727204064.13763: checking for max_fail_percentage 13118 1727204064.13767: done checking for max_fail_percentage 13118 1727204064.13768: checking to see if all hosts have failed and the running result is not ok 13118 1727204064.13768: done checking to see if all hosts have failed 13118 1727204064.13770: getting the remaining hosts for this loop 13118 1727204064.13771: done getting the remaining hosts for this loop 13118 1727204064.13776: getting the next task for host managed-node2 13118 1727204064.13785: done getting next task for host managed-node2 13118 1727204064.13787: ^ task is: TASK: meta (flush_handlers) 13118 1727204064.13789: ^ state is: HOST STATE: block=1, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204064.13793: getting variables 13118 1727204064.13795: in VariableManager get_vars() 13118 1727204064.13819: Calling all_inventory to load vars for managed-node2 13118 1727204064.13822: Calling groups_inventory to load vars for managed-node2 13118 1727204064.13825: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204064.13839: Calling all_plugins_play to load vars for managed-node2 13118 1727204064.13842: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204064.13845: Calling groups_plugins_play to load vars for managed-node2 13118 1727204064.14022: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204064.14409: done with get_vars() 13118 1727204064.14423: done getting variables 13118 1727204064.14495: in VariableManager get_vars() 13118 1727204064.14505: Calling all_inventory to load vars for managed-node2 13118 1727204064.14507: Calling groups_inventory to load vars for managed-node2 13118 1727204064.14510: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204064.14514: Calling all_plugins_play to load vars for managed-node2 13118 1727204064.14516: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204064.14519: Calling groups_plugins_play to load vars for managed-node2 13118 1727204064.15647: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000cc 13118 1727204064.15657: WORKER PROCESS EXITING 13118 1727204064.15984: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204064.16617: done with get_vars() 13118 1727204064.16641: done queuing things up, now waiting for results queue to drain 13118 1727204064.16644: results queue empty 13118 1727204064.16645: checking for any_errors_fatal 13118 1727204064.16648: done checking for any_errors_fatal 13118 1727204064.16649: checking for max_fail_percentage 13118 1727204064.16650: done checking for max_fail_percentage 13118 1727204064.16651: checking to see if all hosts have failed and the running result is not ok 13118 1727204064.16651: done checking to see if all hosts have failed 13118 1727204064.16652: getting the remaining hosts for this loop 13118 1727204064.16653: done getting the remaining hosts for this loop 13118 1727204064.16656: getting the next task for host managed-node2 13118 1727204064.16661: done getting next task for host managed-node2 13118 1727204064.16666: ^ task is: TASK: Include the task 'el_repo_setup.yml' 13118 1727204064.16667: ^ state is: HOST STATE: block=2, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204064.16670: getting variables 13118 1727204064.16671: in VariableManager get_vars() 13118 1727204064.16680: Calling all_inventory to load vars for managed-node2 13118 1727204064.16682: Calling groups_inventory to load vars for managed-node2 13118 1727204064.16685: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204064.16690: Calling all_plugins_play to load vars for managed-node2 13118 1727204064.16692: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204064.16694: Calling groups_plugins_play to load vars for managed-node2 13118 1727204064.17289: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204064.17931: done with get_vars() 13118 1727204064.17942: done getting variables TASK [Include the task 'el_repo_setup.yml'] ************************************ task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tests_bond_nm.yml:11 Tuesday 24 September 2024 14:54:24 -0400 (0:00:01.450) 0:00:01.482 ***** 13118 1727204064.18260: entering _queue_task() for managed-node2/include_tasks 13118 1727204064.18262: Creating lock for include_tasks 13118 1727204064.19251: worker is 1 (out of 1 available) 13118 1727204064.19259: exiting _queue_task() for managed-node2/include_tasks 13118 1727204064.19271: done queuing things up, now waiting for results queue to drain 13118 1727204064.19273: waiting for pending results... 13118 1727204064.19304: running TaskExecutor() for managed-node2/TASK: Include the task 'el_repo_setup.yml' 13118 1727204064.20168: in run() - task 0affcd87-79f5-56a3-0a64-000000000006 13118 1727204064.20190: variable 'ansible_search_path' from source: unknown 13118 1727204064.20236: calling self._execute() 13118 1727204064.20311: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204064.20581: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204064.20598: variable 'omit' from source: magic vars 13118 1727204064.20709: _execute() done 13118 1727204064.20719: dumping result to json 13118 1727204064.20731: done dumping result, returning 13118 1727204064.20744: done running TaskExecutor() for managed-node2/TASK: Include the task 'el_repo_setup.yml' [0affcd87-79f5-56a3-0a64-000000000006] 13118 1727204064.20756: sending task result for task 0affcd87-79f5-56a3-0a64-000000000006 13118 1727204064.20912: no more pending results, returning what we have 13118 1727204064.20917: in VariableManager get_vars() 13118 1727204064.20954: Calling all_inventory to load vars for managed-node2 13118 1727204064.20957: Calling groups_inventory to load vars for managed-node2 13118 1727204064.20960: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204064.20979: Calling all_plugins_play to load vars for managed-node2 13118 1727204064.20983: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204064.20987: Calling groups_plugins_play to load vars for managed-node2 13118 1727204064.21208: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000006 13118 1727204064.21212: WORKER PROCESS EXITING 13118 1727204064.21226: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204064.21409: done with get_vars() 13118 1727204064.21417: variable 'ansible_search_path' from source: unknown 13118 1727204064.21432: we have included files to process 13118 1727204064.21433: generating all_blocks data 13118 1727204064.21435: done generating all_blocks data 13118 1727204064.21436: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml 13118 1727204064.21437: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml 13118 1727204064.21440: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml 13118 1727204064.23112: in VariableManager get_vars() 13118 1727204064.23231: done with get_vars() 13118 1727204064.23908: done processing included file 13118 1727204064.23910: iterating over new_blocks loaded from include file 13118 1727204064.23912: in VariableManager get_vars() 13118 1727204064.23925: done with get_vars() 13118 1727204064.23926: filtering new block on tags 13118 1727204064.23942: done filtering new block on tags 13118 1727204064.23946: in VariableManager get_vars() 13118 1727204064.23981: done with get_vars() 13118 1727204064.23983: filtering new block on tags 13118 1727204064.24000: done filtering new block on tags 13118 1727204064.24229: in VariableManager get_vars() 13118 1727204064.24243: done with get_vars() 13118 1727204064.24245: filtering new block on tags 13118 1727204064.24263: done filtering new block on tags 13118 1727204064.24267: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml for managed-node2 13118 1727204064.24274: extending task lists for all hosts with included blocks 13118 1727204064.24323: done extending task lists 13118 1727204064.24324: done processing included files 13118 1727204064.24325: results queue empty 13118 1727204064.24326: checking for any_errors_fatal 13118 1727204064.24327: done checking for any_errors_fatal 13118 1727204064.24328: checking for max_fail_percentage 13118 1727204064.24606: done checking for max_fail_percentage 13118 1727204064.24608: checking to see if all hosts have failed and the running result is not ok 13118 1727204064.24609: done checking to see if all hosts have failed 13118 1727204064.24610: getting the remaining hosts for this loop 13118 1727204064.24611: done getting the remaining hosts for this loop 13118 1727204064.24614: getting the next task for host managed-node2 13118 1727204064.24619: done getting next task for host managed-node2 13118 1727204064.24621: ^ task is: TASK: Gather the minimum subset of ansible_facts required by the network role test 13118 1727204064.24624: ^ state is: HOST STATE: block=2, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204064.24626: getting variables 13118 1727204064.24627: in VariableManager get_vars() 13118 1727204064.24635: Calling all_inventory to load vars for managed-node2 13118 1727204064.24638: Calling groups_inventory to load vars for managed-node2 13118 1727204064.24640: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204064.24646: Calling all_plugins_play to load vars for managed-node2 13118 1727204064.24648: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204064.24651: Calling groups_plugins_play to load vars for managed-node2 13118 1727204064.25248: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204064.25743: done with get_vars() 13118 1727204064.25837: done getting variables TASK [Gather the minimum subset of ansible_facts required by the network role test] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:3 Tuesday 24 September 2024 14:54:24 -0400 (0:00:00.076) 0:00:01.559 ***** 13118 1727204064.25974: entering _queue_task() for managed-node2/setup 13118 1727204064.26885: worker is 1 (out of 1 available) 13118 1727204064.26898: exiting _queue_task() for managed-node2/setup 13118 1727204064.26909: done queuing things up, now waiting for results queue to drain 13118 1727204064.26910: waiting for pending results... 13118 1727204064.28205: running TaskExecutor() for managed-node2/TASK: Gather the minimum subset of ansible_facts required by the network role test 13118 1727204064.28321: in run() - task 0affcd87-79f5-56a3-0a64-0000000000dd 13118 1727204064.28487: variable 'ansible_search_path' from source: unknown 13118 1727204064.28657: variable 'ansible_search_path' from source: unknown 13118 1727204064.28709: calling self._execute() 13118 1727204064.28988: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204064.28999: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204064.29013: variable 'omit' from source: magic vars 13118 1727204064.31196: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204064.38531: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204064.38610: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204064.38656: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204064.38715: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204064.38748: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204064.38949: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204064.39023: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204064.39055: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204064.39182: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204064.39326: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204064.39590: variable 'ansible_facts' from source: unknown 13118 1727204064.39677: variable 'network_test_required_facts' from source: task vars 13118 1727204064.39792: Evaluated conditional (not ansible_facts.keys() | list | intersect(network_test_required_facts) == network_test_required_facts): True 13118 1727204064.39977: variable 'omit' from source: magic vars 13118 1727204064.40021: variable 'omit' from source: magic vars 13118 1727204064.40099: variable 'omit' from source: magic vars 13118 1727204064.40209: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204064.40243: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204064.40271: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204064.40311: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204064.40414: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204064.40451: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204064.40460: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204064.40472: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204064.40698: Set connection var ansible_timeout to 10 13118 1727204064.40823: Set connection var ansible_pipelining to False 13118 1727204064.40953: Set connection var ansible_connection to ssh 13118 1727204064.40963: Set connection var ansible_shell_executable to /bin/sh 13118 1727204064.40976: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204064.40983: Set connection var ansible_shell_type to sh 13118 1727204064.41012: variable 'ansible_shell_executable' from source: unknown 13118 1727204064.41021: variable 'ansible_connection' from source: unknown 13118 1727204064.41046: variable 'ansible_module_compression' from source: unknown 13118 1727204064.41060: variable 'ansible_shell_type' from source: unknown 13118 1727204064.41068: variable 'ansible_shell_executable' from source: unknown 13118 1727204064.41079: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204064.41086: variable 'ansible_pipelining' from source: unknown 13118 1727204064.41092: variable 'ansible_timeout' from source: unknown 13118 1727204064.41099: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204064.41450: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204064.41726: variable 'omit' from source: magic vars 13118 1727204064.41826: starting attempt loop 13118 1727204064.41884: running the handler 13118 1727204064.41921: _low_level_execute_command(): starting 13118 1727204064.42025: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204064.43393: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204064.43409: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204064.43423: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204064.43443: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204064.43493: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204064.43505: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204064.43519: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204064.43538: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204064.43551: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204064.43572: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204064.43585: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204064.43599: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204064.43616: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204064.43627: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204064.43638: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204064.43652: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204064.43734: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204064.43759: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204064.43784: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204064.43873: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204064.45530: stdout chunk (state=3): >>>/root <<< 13118 1727204064.45729: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204064.45733: stdout chunk (state=3): >>><<< 13118 1727204064.45736: stderr chunk (state=3): >>><<< 13118 1727204064.45859: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204064.45863: _low_level_execute_command(): starting 13118 1727204064.45868: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204064.4576046-13581-235724493291493 `" && echo ansible-tmp-1727204064.4576046-13581-235724493291493="` echo /root/.ansible/tmp/ansible-tmp-1727204064.4576046-13581-235724493291493 `" ) && sleep 0' 13118 1727204064.47289: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204064.47411: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204064.47426: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204064.47444: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204064.47489: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204064.47626: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204064.47643: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204064.47662: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204064.47678: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204064.47691: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204064.47704: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204064.47717: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204064.47737: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204064.47749: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204064.47759: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204064.47775: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204064.47849: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204064.47876: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204064.47900: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204064.47976: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204064.49845: stdout chunk (state=3): >>>ansible-tmp-1727204064.4576046-13581-235724493291493=/root/.ansible/tmp/ansible-tmp-1727204064.4576046-13581-235724493291493 <<< 13118 1727204064.50045: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204064.50050: stdout chunk (state=3): >>><<< 13118 1727204064.50052: stderr chunk (state=3): >>><<< 13118 1727204064.50176: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204064.4576046-13581-235724493291493=/root/.ansible/tmp/ansible-tmp-1727204064.4576046-13581-235724493291493 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204064.50180: variable 'ansible_module_compression' from source: unknown 13118 1727204064.50284: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.setup-ZIP_DEFLATED 13118 1727204064.50287: variable 'ansible_facts' from source: unknown 13118 1727204064.50570: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204064.4576046-13581-235724493291493/AnsiballZ_setup.py 13118 1727204064.51329: Sending initial data 13118 1727204064.51332: Sent initial data (154 bytes) 13118 1727204064.53859: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204064.53884: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204064.53900: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204064.53918: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204064.53970: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204064.54101: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204064.54116: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204064.54137: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204064.54143: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204064.54151: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204064.54159: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204064.54170: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204064.54182: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204064.54190: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204064.54199: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204064.54211: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204064.54341: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204064.54347: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204064.54358: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204064.54485: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204064.56192: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204064.56233: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204064.56274: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmp7ki_w5th /root/.ansible/tmp/ansible-tmp-1727204064.4576046-13581-235724493291493/AnsiballZ_setup.py <<< 13118 1727204064.56309: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204064.59191: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204064.59416: stderr chunk (state=3): >>><<< 13118 1727204064.59420: stdout chunk (state=3): >>><<< 13118 1727204064.59423: done transferring module to remote 13118 1727204064.59425: _low_level_execute_command(): starting 13118 1727204064.59427: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204064.4576046-13581-235724493291493/ /root/.ansible/tmp/ansible-tmp-1727204064.4576046-13581-235724493291493/AnsiballZ_setup.py && sleep 0' 13118 1727204064.61278: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204064.61386: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204064.61408: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204064.61476: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204064.61527: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204064.61579: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204064.61594: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204064.61616: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204064.61628: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204064.61639: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204064.61651: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204064.61666: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204064.61683: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204064.61736: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204064.61748: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204064.61761: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204064.62068: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204064.62085: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204064.62100: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204064.62283: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204064.64071: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204064.64075: stdout chunk (state=3): >>><<< 13118 1727204064.64077: stderr chunk (state=3): >>><<< 13118 1727204064.64176: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204064.64180: _low_level_execute_command(): starting 13118 1727204064.64183: _low_level_execute_command(): executing: /bin/sh -c 'PYTHONVERBOSE=1 /usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204064.4576046-13581-235724493291493/AnsiballZ_setup.py && sleep 0' 13118 1727204064.66055: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204064.66178: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204064.66193: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204064.66211: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204064.66280: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204064.66322: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204064.66383: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204064.66407: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204064.66420: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204064.66431: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204064.66442: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204064.66455: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204064.66472: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204064.66483: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204064.66493: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204064.66510: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204064.66587: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204064.66684: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204064.66699: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204064.66840: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204064.68720: stdout chunk (state=3): >>>import _frozen_importlib # frozen import _imp # builtin <<< 13118 1727204064.68729: stdout chunk (state=3): >>>import '_thread' # import '_warnings' # import '_weakref' # <<< 13118 1727204064.68790: stdout chunk (state=3): >>>import '_io' # <<< 13118 1727204064.68794: stdout chunk (state=3): >>> import 'marshal' # <<< 13118 1727204064.68821: stdout chunk (state=3): >>>import 'posix' # <<< 13118 1727204064.68850: stdout chunk (state=3): >>>import '_frozen_importlib_external' # # installing zipimport hook <<< 13118 1727204064.68898: stdout chunk (state=3): >>>import 'time' # import 'zipimport' # <<< 13118 1727204064.68901: stdout chunk (state=3): >>># installed zipimport hook <<< 13118 1727204064.68952: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/encodings/__init__.py <<< 13118 1727204064.68956: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/encodings/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204064.68974: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/codecs.cpython-39.pyc matches /usr/lib64/python3.9/codecs.py <<< 13118 1727204064.68986: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/codecs.cpython-39.pyc' import '_codecs' # <<< 13118 1727204064.69020: stdout chunk (state=3): >>>import 'codecs' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe83dc0> <<< 13118 1727204064.69049: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/aliases.cpython-39.pyc matches /usr/lib64/python3.9/encodings/aliases.py <<< 13118 1727204064.69066: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/encodings/__pycache__/aliases.cpython-39.pyc' import 'encodings.aliases' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe283a0> <<< 13118 1727204064.69078: stdout chunk (state=3): >>>import 'encodings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe83b20> <<< 13118 1727204064.69095: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/utf_8.cpython-39.pyc matches /usr/lib64/python3.9/encodings/utf_8.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/utf_8.cpython-39.pyc' <<< 13118 1727204064.69110: stdout chunk (state=3): >>>import 'encodings.utf_8' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe83ac0> <<< 13118 1727204064.69131: stdout chunk (state=3): >>>import '_signal' # <<< 13118 1727204064.69152: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/latin_1.cpython-39.pyc matches /usr/lib64/python3.9/encodings/latin_1.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/latin_1.cpython-39.pyc' <<< 13118 1727204064.69171: stdout chunk (state=3): >>>import 'encodings.latin_1' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe28490> <<< 13118 1727204064.69204: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/io.cpython-39.pyc matches /usr/lib64/python3.9/io.py # code object from '/usr/lib64/python3.9/__pycache__/io.cpython-39.pyc' <<< 13118 1727204064.69209: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/abc.py <<< 13118 1727204064.69234: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/abc.cpython-39.pyc' <<< 13118 1727204064.69237: stdout chunk (state=3): >>>import '_abc' # <<< 13118 1727204064.69241: stdout chunk (state=3): >>>import 'abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe28940> <<< 13118 1727204064.69256: stdout chunk (state=3): >>>import 'io' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe28670> <<< 13118 1727204064.69287: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/site.cpython-39.pyc matches /usr/lib64/python3.9/site.py <<< 13118 1727204064.69298: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/site.cpython-39.pyc' <<< 13118 1727204064.69320: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/os.cpython-39.pyc matches /usr/lib64/python3.9/os.py <<< 13118 1727204064.69335: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/os.cpython-39.pyc' <<< 13118 1727204064.69355: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/stat.cpython-39.pyc matches /usr/lib64/python3.9/stat.py <<< 13118 1727204064.69374: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/stat.cpython-39.pyc' <<< 13118 1727204064.69393: stdout chunk (state=3): >>>import '_stat' # import 'stat' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fbcf190> <<< 13118 1727204064.69413: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_collections_abc.cpython-39.pyc matches /usr/lib64/python3.9/_collections_abc.py <<< 13118 1727204064.69430: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/_collections_abc.cpython-39.pyc' <<< 13118 1727204064.69504: stdout chunk (state=3): >>>import '_collections_abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fbcf220> <<< 13118 1727204064.69534: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/posixpath.cpython-39.pyc matches /usr/lib64/python3.9/posixpath.py <<< 13118 1727204064.69537: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/posixpath.cpython-39.pyc' <<< 13118 1727204064.69581: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/genericpath.cpython-39.pyc matches /usr/lib64/python3.9/genericpath.py # code object from '/usr/lib64/python3.9/__pycache__/genericpath.cpython-39.pyc' <<< 13118 1727204064.69585: stdout chunk (state=3): >>>import 'genericpath' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fbf2850> import 'posixpath' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fbcf940> <<< 13118 1727204064.69597: stdout chunk (state=3): >>>import 'os' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe40880> <<< 13118 1727204064.69624: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_sitebuiltins.cpython-39.pyc matches /usr/lib64/python3.9/_sitebuiltins.py # code object from '/usr/lib64/python3.9/__pycache__/_sitebuiltins.cpython-39.pyc' <<< 13118 1727204064.69627: stdout chunk (state=3): >>>import '_sitebuiltins' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fbc8d90> <<< 13118 1727204064.69688: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_bootlocale.cpython-39.pyc matches /usr/lib64/python3.9/_bootlocale.py <<< 13118 1727204064.69691: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/_bootlocale.cpython-39.pyc' import '_locale' # <<< 13118 1727204064.69693: stdout chunk (state=3): >>>import '_bootlocale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fbf2d90> <<< 13118 1727204064.69744: stdout chunk (state=3): >>>import 'site' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe28970> <<< 13118 1727204064.69774: stdout chunk (state=3): >>>Python 3.9.19 (main, Aug 23 2024, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-2)] on linux Type "help", "copyright", "credits" or "license" for more information. <<< 13118 1727204064.70098: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/base64.cpython-39.pyc matches /usr/lib64/python3.9/base64.py <<< 13118 1727204064.70121: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/base64.cpython-39.pyc' <<< 13118 1727204064.70145: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/re.cpython-39.pyc matches /usr/lib64/python3.9/re.py <<< 13118 1727204064.70148: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/re.cpython-39.pyc' <<< 13118 1727204064.70169: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/enum.cpython-39.pyc matches /usr/lib64/python3.9/enum.py <<< 13118 1727204064.70183: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/enum.cpython-39.pyc' <<< 13118 1727204064.70206: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/types.cpython-39.pyc matches /usr/lib64/python3.9/types.py <<< 13118 1727204064.70209: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/types.cpython-39.pyc' <<< 13118 1727204064.70220: stdout chunk (state=3): >>>import 'types' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb6ef10> <<< 13118 1727204064.70269: stdout chunk (state=3): >>>import 'enum' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb740a0> <<< 13118 1727204064.70295: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/sre_compile.cpython-39.pyc matches /usr/lib64/python3.9/sre_compile.py <<< 13118 1727204064.70298: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/sre_compile.cpython-39.pyc' <<< 13118 1727204064.70314: stdout chunk (state=3): >>>import '_sre' # <<< 13118 1727204064.70336: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/sre_parse.cpython-39.pyc matches /usr/lib64/python3.9/sre_parse.py <<< 13118 1727204064.70339: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/sre_parse.cpython-39.pyc' <<< 13118 1727204064.70367: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/sre_constants.cpython-39.pyc matches /usr/lib64/python3.9/sre_constants.py # code object from '/usr/lib64/python3.9/__pycache__/sre_constants.cpython-39.pyc' <<< 13118 1727204064.70390: stdout chunk (state=3): >>>import 'sre_constants' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb675b0> <<< 13118 1727204064.70412: stdout chunk (state=3): >>>import 'sre_parse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb6f6a0> <<< 13118 1727204064.70419: stdout chunk (state=3): >>>import 'sre_compile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb6e3d0> <<< 13118 1727204064.70437: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/functools.cpython-39.pyc matches /usr/lib64/python3.9/functools.py <<< 13118 1727204064.70512: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/functools.cpython-39.pyc' <<< 13118 1727204064.70526: stdout chunk (state=3): >>># /usr/lib64/python3.9/collections/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/collections/__init__.py <<< 13118 1727204064.70573: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/collections/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204064.70607: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/heapq.cpython-39.pyc matches /usr/lib64/python3.9/heapq.py # code object from '/usr/lib64/python3.9/__pycache__/heapq.cpython-39.pyc' <<< 13118 1727204064.70638: stdout chunk (state=3): >>># extension module '_heapq' loaded from '/usr/lib64/python3.9/lib-dynload/_heapq.cpython-39-x86_64-linux-gnu.so' # extension module '_heapq' executed from '/usr/lib64/python3.9/lib-dynload/_heapq.cpython-39-x86_64-linux-gnu.so' import '_heapq' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211fa55eb0> import 'heapq' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa559a0> import 'itertools' # <<< 13118 1727204064.70674: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/keyword.cpython-39.pyc matches /usr/lib64/python3.9/keyword.py # code object from '/usr/lib64/python3.9/__pycache__/keyword.cpython-39.pyc' import 'keyword' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa55fa0> <<< 13118 1727204064.70729: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/operator.cpython-39.pyc matches /usr/lib64/python3.9/operator.py # code object from '/usr/lib64/python3.9/__pycache__/operator.cpython-39.pyc' <<< 13118 1727204064.70753: stdout chunk (state=3): >>>import '_operator' # import 'operator' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa55df0> <<< 13118 1727204064.70756: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/reprlib.cpython-39.pyc matches /usr/lib64/python3.9/reprlib.py # code object from '/usr/lib64/python3.9/__pycache__/reprlib.cpython-39.pyc' import 'reprlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa65160> import '_collections' # <<< 13118 1727204064.70809: stdout chunk (state=3): >>>import 'collections' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb49e20> import '_functools' # <<< 13118 1727204064.70833: stdout chunk (state=3): >>>import 'functools' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb41700> <<< 13118 1727204064.70905: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/copyreg.cpython-39.pyc matches /usr/lib64/python3.9/copyreg.py # code object from '/usr/lib64/python3.9/__pycache__/copyreg.cpython-39.pyc' import 'copyreg' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb55760> import 're' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb75eb0> <<< 13118 1727204064.70942: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/struct.cpython-39.pyc matches /usr/lib64/python3.9/struct.py # code object from '/usr/lib64/python3.9/__pycache__/struct.cpython-39.pyc' <<< 13118 1727204064.70983: stdout chunk (state=3): >>># extension module '_struct' loaded from '/usr/lib64/python3.9/lib-dynload/_struct.cpython-39-x86_64-linux-gnu.so' # extension module '_struct' executed from '/usr/lib64/python3.9/lib-dynload/_struct.cpython-39-x86_64-linux-gnu.so' import '_struct' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211fa65d60> import 'struct' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb49340> <<< 13118 1727204064.71015: stdout chunk (state=3): >>># extension module 'binascii' loaded from '/usr/lib64/python3.9/lib-dynload/binascii.cpython-39-x86_64-linux-gnu.so' # extension module 'binascii' executed from '/usr/lib64/python3.9/lib-dynload/binascii.cpython-39-x86_64-linux-gnu.so' import 'binascii' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211fb55370> import 'base64' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb7ba60> <<< 13118 1727204064.71067: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/runpy.cpython-39.pyc matches /usr/lib64/python3.9/runpy.py # code object from '/usr/lib64/python3.9/__pycache__/runpy.cpython-39.pyc' <<< 13118 1727204064.71116: stdout chunk (state=3): >>># /usr/lib64/python3.9/importlib/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/importlib/__init__.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204064.71121: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/warnings.cpython-39.pyc matches /usr/lib64/python3.9/warnings.py # code object from '/usr/lib64/python3.9/__pycache__/warnings.cpython-39.pyc' import 'warnings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa65f40> import 'importlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa65e80> <<< 13118 1727204064.71174: stdout chunk (state=3): >>># /usr/lib64/python3.9/importlib/__pycache__/machinery.cpython-39.pyc matches /usr/lib64/python3.9/importlib/machinery.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/machinery.cpython-39.pyc' import 'importlib.machinery' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa65df0> # /usr/lib64/python3.9/importlib/__pycache__/util.cpython-39.pyc matches /usr/lib64/python3.9/importlib/util.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/util.cpython-39.pyc' <<< 13118 1727204064.71205: stdout chunk (state=3): >>># /usr/lib64/python3.9/importlib/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/importlib/abc.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/abc.cpython-39.pyc' <<< 13118 1727204064.71210: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/typing.cpython-39.pyc matches /usr/lib64/python3.9/typing.py <<< 13118 1727204064.71262: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/typing.cpython-39.pyc' <<< 13118 1727204064.71304: stdout chunk (state=3): >>># /usr/lib64/python3.9/collections/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/collections/abc.py # code object from '/usr/lib64/python3.9/collections/__pycache__/abc.cpython-39.pyc' import 'collections.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa39460> <<< 13118 1727204064.71308: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/contextlib.cpython-39.pyc matches /usr/lib64/python3.9/contextlib.py # code object from '/usr/lib64/python3.9/__pycache__/contextlib.cpython-39.pyc' <<< 13118 1727204064.71340: stdout chunk (state=3): >>>import 'contextlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa39550> <<< 13118 1727204064.71454: stdout chunk (state=3): >>>import 'typing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa170d0> <<< 13118 1727204064.71517: stdout chunk (state=3): >>>import 'importlib.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa68b20> import 'importlib.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa684c0> <<< 13118 1727204064.71534: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/pkgutil.cpython-39.pyc matches /usr/lib64/python3.9/pkgutil.py # code object from '/usr/lib64/python3.9/__pycache__/pkgutil.cpython-39.pyc' <<< 13118 1727204064.71594: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/weakref.cpython-39.pyc matches /usr/lib64/python3.9/weakref.py <<< 13118 1727204064.71613: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/weakref.cpython-39.pyc' <<< 13118 1727204064.71636: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_weakrefset.cpython-39.pyc matches /usr/lib64/python3.9/_weakrefset.py # code object from '/usr/lib64/python3.9/__pycache__/_weakrefset.cpython-39.pyc' import '_weakrefset' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f95f2b0> <<< 13118 1727204064.71650: stdout chunk (state=3): >>>import 'weakref' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa24d60> <<< 13118 1727204064.71694: stdout chunk (state=3): >>>import 'pkgutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa68fa0> import 'runpy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb7b0d0> <<< 13118 1727204064.71734: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/shutil.cpython-39.pyc matches /usr/lib64/python3.9/shutil.py <<< 13118 1727204064.71777: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/shutil.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/fnmatch.cpython-39.pyc matches /usr/lib64/python3.9/fnmatch.py # code object from '/usr/lib64/python3.9/__pycache__/fnmatch.cpython-39.pyc' <<< 13118 1727204064.71805: stdout chunk (state=3): >>>import 'fnmatch' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f96ebe0> import 'errno' # <<< 13118 1727204064.71855: stdout chunk (state=3): >>># extension module 'zlib' loaded from '/usr/lib64/python3.9/lib-dynload/zlib.cpython-39-x86_64-linux-gnu.so' # extension module 'zlib' executed from '/usr/lib64/python3.9/lib-dynload/zlib.cpython-39-x86_64-linux-gnu.so' import 'zlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f96ef10> <<< 13118 1727204064.71888: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/bz2.cpython-39.pyc matches /usr/lib64/python3.9/bz2.py # code object from '/usr/lib64/python3.9/__pycache__/bz2.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/_compression.cpython-39.pyc matches /usr/lib64/python3.9/_compression.py # code object from '/usr/lib64/python3.9/__pycache__/_compression.cpython-39.pyc' import '_compression' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f982820> <<< 13118 1727204064.71901: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/threading.cpython-39.pyc matches /usr/lib64/python3.9/threading.py <<< 13118 1727204064.71925: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/threading.cpython-39.pyc' <<< 13118 1727204064.71953: stdout chunk (state=3): >>>import 'threading' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f982d60> <<< 13118 1727204064.71996: stdout chunk (state=3): >>># extension module '_bz2' loaded from '/usr/lib64/python3.9/lib-dynload/_bz2.cpython-39-x86_64-linux-gnu.so' # extension module '_bz2' executed from '/usr/lib64/python3.9/lib-dynload/_bz2.cpython-39-x86_64-linux-gnu.so' import '_bz2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f90f490> import 'bz2' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f96ef40> <<< 13118 1727204064.72024: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/lzma.cpython-39.pyc matches /usr/lib64/python3.9/lzma.py # code object from '/usr/lib64/python3.9/__pycache__/lzma.cpython-39.pyc' <<< 13118 1727204064.72084: stdout chunk (state=3): >>># extension module '_lzma' loaded from '/usr/lib64/python3.9/lib-dynload/_lzma.cpython-39-x86_64-linux-gnu.so' # extension module '_lzma' executed from '/usr/lib64/python3.9/lib-dynload/_lzma.cpython-39-x86_64-linux-gnu.so' import '_lzma' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f91f370> import 'lzma' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f9826a0> import 'pwd' # <<< 13118 1727204064.72112: stdout chunk (state=3): >>># extension module 'grp' loaded from '/usr/lib64/python3.9/lib-dynload/grp.cpython-39-x86_64-linux-gnu.so' # extension module 'grp' executed from '/usr/lib64/python3.9/lib-dynload/grp.cpython-39-x86_64-linux-gnu.so' import 'grp' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f91f430> <<< 13118 1727204064.72149: stdout chunk (state=3): >>>import 'shutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa65ac0> <<< 13118 1727204064.72194: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/tempfile.cpython-39.pyc matches /usr/lib64/python3.9/tempfile.py <<< 13118 1727204064.72217: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/tempfile.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/random.cpython-39.pyc matches /usr/lib64/python3.9/random.py # code object from '/usr/lib64/python3.9/__pycache__/random.cpython-39.pyc' <<< 13118 1727204064.72251: stdout chunk (state=3): >>># extension module 'math' loaded from '/usr/lib64/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so' # extension module 'math' executed from '/usr/lib64/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so' import 'math' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f93b790> <<< 13118 1727204064.72301: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/bisect.cpython-39.pyc matches /usr/lib64/python3.9/bisect.py # code object from '/usr/lib64/python3.9/__pycache__/bisect.cpython-39.pyc' <<< 13118 1727204064.72343: stdout chunk (state=3): >>># extension module '_bisect' loaded from '/usr/lib64/python3.9/lib-dynload/_bisect.cpython-39-x86_64-linux-gnu.so' # extension module '_bisect' executed from '/usr/lib64/python3.9/lib-dynload/_bisect.cpython-39-x86_64-linux-gnu.so' import '_bisect' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f93ba60> import 'bisect' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f93b850> # extension module '_random' loaded from '/usr/lib64/python3.9/lib-dynload/_random.cpython-39-x86_64-linux-gnu.so' # extension module '_random' executed from '/usr/lib64/python3.9/lib-dynload/_random.cpython-39-x86_64-linux-gnu.so' import '_random' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f93b940> <<< 13118 1727204064.72357: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/hashlib.cpython-39.pyc matches /usr/lib64/python3.9/hashlib.py # code object from '/usr/lib64/python3.9/__pycache__/hashlib.cpython-39.pyc' <<< 13118 1727204064.72545: stdout chunk (state=3): >>># extension module '_hashlib' loaded from '/usr/lib64/python3.9/lib-dynload/_hashlib.cpython-39-x86_64-linux-gnu.so' # extension module '_hashlib' executed from '/usr/lib64/python3.9/lib-dynload/_hashlib.cpython-39-x86_64-linux-gnu.so' import '_hashlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f93bd90> <<< 13118 1727204064.72594: stdout chunk (state=3): >>># extension module '_blake2' loaded from '/usr/lib64/python3.9/lib-dynload/_blake2.cpython-39-x86_64-linux-gnu.so' # extension module '_blake2' executed from '/usr/lib64/python3.9/lib-dynload/_blake2.cpython-39-x86_64-linux-gnu.so' import '_blake2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f9462e0> <<< 13118 1727204064.72607: stdout chunk (state=3): >>>import 'hashlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f93b9d0> import 'random' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f92fb20> <<< 13118 1727204064.72647: stdout chunk (state=3): >>>import 'tempfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa656a0> # /usr/lib64/python3.9/__pycache__/zipfile.cpython-39.pyc matches /usr/lib64/python3.9/zipfile.py <<< 13118 1727204064.72706: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/zipfile.cpython-39.pyc' <<< 13118 1727204064.72745: stdout chunk (state=3): >>>import 'zipfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f93bb80> <<< 13118 1727204064.72890: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/encodings/cp437.pyc' <<< 13118 1727204064.72903: stdout chunk (state=3): >>>import 'encodings.cp437' # <_frozen_importlib_external.SourcelessFileLoader object at 0x7f211f85d760> <<< 13118 1727204064.73152: stdout chunk (state=3): >>># zipimport: found 103 names in '/tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip' # zipimport: zlib available <<< 13118 1727204064.73255: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.73300: stdout chunk (state=3): >>>import ansible # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/__init__.py <<< 13118 1727204064.73330: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13118 1727204064.73348: stdout chunk (state=3): >>>import ansible.module_utils # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/__init__.py <<< 13118 1727204064.73351: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.74582: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.75551: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/__future__.cpython-39.pyc matches /usr/lib64/python3.9/__future__.py # code object from '/usr/lib64/python3.9/__pycache__/__future__.cpython-39.pyc' import '__future__' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f79b8b0> <<< 13118 1727204064.75585: stdout chunk (state=3): >>># /usr/lib64/python3.9/json/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/json/__init__.py # code object from '/usr/lib64/python3.9/json/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204064.75620: stdout chunk (state=3): >>># /usr/lib64/python3.9/json/__pycache__/decoder.cpython-39.pyc matches /usr/lib64/python3.9/json/decoder.py # code object from '/usr/lib64/python3.9/json/__pycache__/decoder.cpython-39.pyc' <<< 13118 1727204064.75624: stdout chunk (state=3): >>># /usr/lib64/python3.9/json/__pycache__/scanner.cpython-39.pyc matches /usr/lib64/python3.9/json/scanner.py # code object from '/usr/lib64/python3.9/json/__pycache__/scanner.cpython-39.pyc' # extension module '_json' loaded from '/usr/lib64/python3.9/lib-dynload/_json.cpython-39-x86_64-linux-gnu.so' # extension module '_json' executed from '/usr/lib64/python3.9/lib-dynload/_json.cpython-39-x86_64-linux-gnu.so' import '_json' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f79b160> <<< 13118 1727204064.75654: stdout chunk (state=3): >>>import 'json.scanner' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f79b280> <<< 13118 1727204064.75715: stdout chunk (state=3): >>>import 'json.decoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f79b5e0> # /usr/lib64/python3.9/json/__pycache__/encoder.cpython-39.pyc matches /usr/lib64/python3.9/json/encoder.py # code object from '/usr/lib64/python3.9/json/__pycache__/encoder.cpython-39.pyc' <<< 13118 1727204064.75768: stdout chunk (state=3): >>>import 'json.encoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f79b4f0> import 'json' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f79be20> import 'atexit' # <<< 13118 1727204064.75813: stdout chunk (state=3): >>># extension module 'fcntl' loaded from '/usr/lib64/python3.9/lib-dynload/fcntl.cpython-39-x86_64-linux-gnu.so' # extension module 'fcntl' executed from '/usr/lib64/python3.9/lib-dynload/fcntl.cpython-39-x86_64-linux-gnu.so' import 'fcntl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f79b580> <<< 13118 1727204064.75838: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/locale.cpython-39.pyc matches /usr/lib64/python3.9/locale.py <<< 13118 1727204064.75852: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/locale.cpython-39.pyc' <<< 13118 1727204064.75894: stdout chunk (state=3): >>>import 'locale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f79b100> <<< 13118 1727204064.75915: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/platform.cpython-39.pyc matches /usr/lib64/python3.9/platform.py <<< 13118 1727204064.75957: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/platform.cpython-39.pyc' <<< 13118 1727204064.75980: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/subprocess.cpython-39.pyc matches /usr/lib64/python3.9/subprocess.py <<< 13118 1727204064.75993: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/subprocess.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/signal.cpython-39.pyc matches /usr/lib64/python3.9/signal.py # code object from '/usr/lib64/python3.9/__pycache__/signal.cpython-39.pyc' <<< 13118 1727204064.76077: stdout chunk (state=3): >>>import 'signal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f730040> <<< 13118 1727204064.76113: stdout chunk (state=3): >>># extension module '_posixsubprocess' loaded from '/usr/lib64/python3.9/lib-dynload/_posixsubprocess.cpython-39-x86_64-linux-gnu.so' # extension module '_posixsubprocess' executed from '/usr/lib64/python3.9/lib-dynload/_posixsubprocess.cpython-39-x86_64-linux-gnu.so' import '_posixsubprocess' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1493d0> <<< 13118 1727204064.76167: stdout chunk (state=3): >>># extension module 'select' loaded from '/usr/lib64/python3.9/lib-dynload/select.cpython-39-x86_64-linux-gnu.so' # extension module 'select' executed from '/usr/lib64/python3.9/lib-dynload/select.cpython-39-x86_64-linux-gnu.so' import 'select' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1490d0> <<< 13118 1727204064.76187: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/selectors.cpython-39.pyc matches /usr/lib64/python3.9/selectors.py # code object from '/usr/lib64/python3.9/__pycache__/selectors.cpython-39.pyc' <<< 13118 1727204064.76220: stdout chunk (state=3): >>>import 'selectors' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f149d30> <<< 13118 1727204064.76235: stdout chunk (state=3): >>>import 'subprocess' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f782d90> <<< 13118 1727204064.76403: stdout chunk (state=3): >>>import 'platform' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7823a0> <<< 13118 1727204064.76419: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/shlex.cpython-39.pyc matches /usr/lib64/python3.9/shlex.py # code object from '/usr/lib64/python3.9/__pycache__/shlex.cpython-39.pyc' <<< 13118 1727204064.76445: stdout chunk (state=3): >>>import 'shlex' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f782f40> <<< 13118 1727204064.76476: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/traceback.cpython-39.pyc matches /usr/lib64/python3.9/traceback.py <<< 13118 1727204064.76479: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/traceback.cpython-39.pyc' <<< 13118 1727204064.76516: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/linecache.cpython-39.pyc matches /usr/lib64/python3.9/linecache.py # code object from '/usr/lib64/python3.9/__pycache__/linecache.cpython-39.pyc' <<< 13118 1727204064.76548: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/tokenize.cpython-39.pyc matches /usr/lib64/python3.9/tokenize.py # code object from '/usr/lib64/python3.9/__pycache__/tokenize.cpython-39.pyc' <<< 13118 1727204064.76575: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/token.cpython-39.pyc matches /usr/lib64/python3.9/token.py # code object from '/usr/lib64/python3.9/__pycache__/token.cpython-39.pyc' import 'token' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f85da90> <<< 13118 1727204064.76656: stdout chunk (state=3): >>>import 'tokenize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f75adc0> import 'linecache' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f75a490> import 'traceback' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f798a90> <<< 13118 1727204064.76691: stdout chunk (state=3): >>># extension module 'syslog' loaded from '/usr/lib64/python3.9/lib-dynload/syslog.cpython-39-x86_64-linux-gnu.so' # extension module 'syslog' executed from '/usr/lib64/python3.9/lib-dynload/syslog.cpython-39-x86_64-linux-gnu.so' import 'syslog' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f75a5b0> <<< 13118 1727204064.76736: stdout chunk (state=3): >>># /usr/lib64/python3.9/site-packages/systemd/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/__init__.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/__init__.cpython-39.pyc' import 'systemd' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f75a5e0> <<< 13118 1727204064.76773: stdout chunk (state=3): >>># /usr/lib64/python3.9/site-packages/systemd/__pycache__/journal.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/journal.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/journal.cpython-39.pyc' <<< 13118 1727204064.76787: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/datetime.cpython-39.pyc matches /usr/lib64/python3.9/datetime.py <<< 13118 1727204064.76817: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/datetime.cpython-39.pyc' <<< 13118 1727204064.76887: stdout chunk (state=3): >>># extension module '_datetime' loaded from '/usr/lib64/python3.9/lib-dynload/_datetime.cpython-39-x86_64-linux-gnu.so' # extension module '_datetime' executed from '/usr/lib64/python3.9/lib-dynload/_datetime.cpython-39-x86_64-linux-gnu.so' import '_datetime' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1b4f70> import 'datetime' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7e52e0> <<< 13118 1727204064.76920: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/uuid.cpython-39.pyc matches /usr/lib64/python3.9/uuid.py # code object from '/usr/lib64/python3.9/__pycache__/uuid.cpython-39.pyc' <<< 13118 1727204064.76983: stdout chunk (state=3): >>># extension module '_uuid' loaded from '/usr/lib64/python3.9/lib-dynload/_uuid.cpython-39-x86_64-linux-gnu.so' # extension module '_uuid' executed from '/usr/lib64/python3.9/lib-dynload/_uuid.cpython-39-x86_64-linux-gnu.so' import '_uuid' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1b17f0> import 'uuid' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7e5460> <<< 13118 1727204064.77007: stdout chunk (state=3): >>># /usr/lib64/python3.9/logging/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/logging/__init__.py <<< 13118 1727204064.77062: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/logging/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204064.77086: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/string.cpython-39.pyc matches /usr/lib64/python3.9/string.py # code object from '/usr/lib64/python3.9/__pycache__/string.cpython-39.pyc' import '_string' # <<< 13118 1727204064.77146: stdout chunk (state=3): >>>import 'string' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7e5c40> <<< 13118 1727204064.77278: stdout chunk (state=3): >>>import 'logging' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f1b1790> <<< 13118 1727204064.77368: stdout chunk (state=3): >>># extension module 'systemd._journal' loaded from '/usr/lib64/python3.9/site-packages/systemd/_journal.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._journal' executed from '/usr/lib64/python3.9/site-packages/systemd/_journal.cpython-39-x86_64-linux-gnu.so' import 'systemd._journal' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f7e5130> <<< 13118 1727204064.77408: stdout chunk (state=3): >>># extension module 'systemd._reader' loaded from '/usr/lib64/python3.9/site-packages/systemd/_reader.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._reader' executed from '/usr/lib64/python3.9/site-packages/systemd/_reader.cpython-39-x86_64-linux-gnu.so' import 'systemd._reader' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f7e5670> <<< 13118 1727204064.77460: stdout chunk (state=3): >>># extension module 'systemd.id128' loaded from '/usr/lib64/python3.9/site-packages/systemd/id128.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd.id128' executed from '/usr/lib64/python3.9/site-packages/systemd/id128.cpython-39-x86_64-linux-gnu.so' import 'systemd.id128' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f7e5730> import 'systemd.journal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7dc9a0> <<< 13118 1727204064.77495: stdout chunk (state=3): >>># /usr/lib64/python3.9/site-packages/systemd/__pycache__/daemon.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/daemon.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/daemon.cpython-39.pyc' <<< 13118 1727204064.77534: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/socket.cpython-39.pyc matches /usr/lib64/python3.9/socket.py <<< 13118 1727204064.77538: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/socket.cpython-39.pyc' <<< 13118 1727204064.77577: stdout chunk (state=3): >>># extension module '_socket' loaded from '/usr/lib64/python3.9/lib-dynload/_socket.cpython-39-x86_64-linux-gnu.so' # extension module '_socket' executed from '/usr/lib64/python3.9/lib-dynload/_socket.cpython-39-x86_64-linux-gnu.so' import '_socket' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1a78e0> <<< 13118 1727204064.77758: stdout chunk (state=3): >>># extension module 'array' loaded from '/usr/lib64/python3.9/lib-dynload/array.cpython-39-x86_64-linux-gnu.so' # extension module 'array' executed from '/usr/lib64/python3.9/lib-dynload/array.cpython-39-x86_64-linux-gnu.so' import 'array' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1c5c70> import 'socket' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f1b0520> <<< 13118 1727204064.77827: stdout chunk (state=3): >>># extension module 'systemd._daemon' loaded from '/usr/lib64/python3.9/site-packages/systemd/_daemon.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._daemon' executed from '/usr/lib64/python3.9/site-packages/systemd/_daemon.cpython-39-x86_64-linux-gnu.so' import 'systemd._daemon' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1a7e80> import 'systemd.daemon' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f1b0940> <<< 13118 1727204064.77849: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.77867: stdout chunk (state=3): >>># zipimport: zlib available import ansible.module_utils.compat # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/compat/__init__.py # zipimport: zlib available <<< 13118 1727204064.77932: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.78023: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13118 1727204064.78070: stdout chunk (state=3): >>>import ansible.module_utils.common # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.text # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/text/__init__.py <<< 13118 1727204064.78084: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.78171: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.78268: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.78716: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.79204: stdout chunk (state=3): >>>import ansible.module_utils.six # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/six/__init__.py import 'ansible.module_utils.six.moves' # import 'ansible.module_utils.six.moves.collections_abc' # import ansible.module_utils.common.text.converters # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/text/converters.py <<< 13118 1727204064.79231: stdout chunk (state=3): >>># /usr/lib64/python3.9/ctypes/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/ctypes/__init__.py # code object from '/usr/lib64/python3.9/ctypes/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204064.79300: stdout chunk (state=3): >>># extension module '_ctypes' loaded from '/usr/lib64/python3.9/lib-dynload/_ctypes.cpython-39-x86_64-linux-gnu.so' # extension module '_ctypes' executed from '/usr/lib64/python3.9/lib-dynload/_ctypes.cpython-39-x86_64-linux-gnu.so' import '_ctypes' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1c0790> <<< 13118 1727204064.79377: stdout chunk (state=3): >>># /usr/lib64/python3.9/ctypes/__pycache__/_endian.cpython-39.pyc matches /usr/lib64/python3.9/ctypes/_endian.py # code object from '/usr/lib64/python3.9/ctypes/__pycache__/_endian.cpython-39.pyc' import 'ctypes._endian' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f1ff850> <<< 13118 1727204064.79380: stdout chunk (state=3): >>>import 'ctypes' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed65fa0> <<< 13118 1727204064.79471: stdout chunk (state=3): >>>import ansible.module_utils.compat.selinux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/compat/selinux.py # zipimport: zlib available <<< 13118 1727204064.79475: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.79487: stdout chunk (state=3): >>>import ansible.module_utils._text # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/_text.py # zipimport: zlib available <<< 13118 1727204064.79602: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.79755: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/copy.cpython-39.pyc matches /usr/lib64/python3.9/copy.py # code object from '/usr/lib64/python3.9/__pycache__/copy.cpython-39.pyc' <<< 13118 1727204064.79773: stdout chunk (state=3): >>>import 'copy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f762310> # zipimport: zlib available <<< 13118 1727204064.80159: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.80530: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.80584: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.80653: stdout chunk (state=3): >>>import ansible.module_utils.common.collections # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/collections.py # zipimport: zlib available <<< 13118 1727204064.80727: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.80743: stdout chunk (state=3): >>>import ansible.module_utils.common.warnings # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/warnings.py # zipimport: zlib available <<< 13118 1727204064.80791: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.80900: stdout chunk (state=3): >>>import ansible.module_utils.errors # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/errors.py # zipimport: zlib available <<< 13118 1727204064.80904: stdout chunk (state=3): >>># zipimport: zlib available import ansible.module_utils.parsing # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/parsing/__init__.py <<< 13118 1727204064.80915: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.80945: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.80991: stdout chunk (state=3): >>>import ansible.module_utils.parsing.convert_bool # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/parsing/convert_bool.py # zipimport: zlib available <<< 13118 1727204064.81179: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.81369: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/ast.cpython-39.pyc matches /usr/lib64/python3.9/ast.py <<< 13118 1727204064.81402: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/ast.cpython-39.pyc' import '_ast' # <<< 13118 1727204064.81483: stdout chunk (state=3): >>>import 'ast' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7a1ca0> # zipimport: zlib available <<< 13118 1727204064.81549: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.81643: stdout chunk (state=3): >>>import ansible.module_utils.common.text.formatters # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/text/formatters.py import ansible.module_utils.common.validation # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/validation.py import ansible.module_utils.common.parameters # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/parameters.py <<< 13118 1727204064.81651: stdout chunk (state=3): >>>import ansible.module_utils.common.arg_spec # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/arg_spec.py <<< 13118 1727204064.81678: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.81691: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.81731: stdout chunk (state=3): >>>import ansible.module_utils.common.locale # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/locale.py # zipimport: zlib available <<< 13118 1727204064.81780: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.81811: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.81905: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.81969: stdout chunk (state=3): >>># /usr/lib64/python3.9/site-packages/selinux/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/selinux/__init__.py <<< 13118 1727204064.82000: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/site-packages/selinux/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204064.82074: stdout chunk (state=3): >>># extension module 'selinux._selinux' loaded from '/usr/lib64/python3.9/site-packages/selinux/_selinux.cpython-39-x86_64-linux-gnu.so' # extension module 'selinux._selinux' executed from '/usr/lib64/python3.9/site-packages/selinux/_selinux.cpython-39-x86_64-linux-gnu.so' import 'selinux._selinux' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1e63a0> <<< 13118 1727204064.82167: stdout chunk (state=3): >>>import 'selinux' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7a1bb0> <<< 13118 1727204064.82218: stdout chunk (state=3): >>>import ansible.module_utils.common.file # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/file.py import ansible.module_utils.common.process # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/process.py <<< 13118 1727204064.82221: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.82270: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.82322: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.82354: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.82412: stdout chunk (state=3): >>># /usr/lib/python3.9/site-packages/__pycache__/distro.cpython-39.pyc matches /usr/lib/python3.9/site-packages/distro.py <<< 13118 1727204064.82416: stdout chunk (state=3): >>># code object from '/usr/lib/python3.9/site-packages/__pycache__/distro.cpython-39.pyc' <<< 13118 1727204064.82431: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/argparse.cpython-39.pyc matches /usr/lib64/python3.9/argparse.py <<< 13118 1727204064.82478: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/argparse.cpython-39.pyc' <<< 13118 1727204064.82507: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/gettext.cpython-39.pyc matches /usr/lib64/python3.9/gettext.py <<< 13118 1727204064.82510: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/gettext.cpython-39.pyc' <<< 13118 1727204064.82591: stdout chunk (state=3): >>>import 'gettext' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f1c22b0> <<< 13118 1727204064.82629: stdout chunk (state=3): >>>import 'argparse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f76fb80> <<< 13118 1727204064.82692: stdout chunk (state=3): >>>import 'distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ebd7eb0> # destroy ansible.module_utils.distro import ansible.module_utils.distro # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/distro/__init__.py # zipimport: zlib available <<< 13118 1727204064.82738: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.82751: stdout chunk (state=3): >>>import ansible.module_utils.common._utils # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/_utils.py import ansible.module_utils.common.sys_info # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/sys_info.py <<< 13118 1727204064.82852: stdout chunk (state=3): >>>import ansible.module_utils.basic # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/basic.py # zipimport: zlib available # zipimport: zlib available <<< 13118 1727204064.82870: stdout chunk (state=3): >>>import ansible.modules # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/modules/__init__.py # zipimport: zlib available <<< 13118 1727204064.82922: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.82996: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.83000: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.83012: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.83047: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.83081: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.83109: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.83150: stdout chunk (state=3): >>>import ansible.module_utils.facts.namespace # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/namespace.py # zipimport: zlib available <<< 13118 1727204064.83216: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.83299: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.83319: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.83336: stdout chunk (state=3): >>>import ansible.module_utils.compat.typing # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/compat/typing.py # zipimport: zlib available <<< 13118 1727204064.83483: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.83616: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.83649: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.83714: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/__init__.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204064.83746: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/context.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/context.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/context.cpython-39.pyc' <<< 13118 1727204064.83760: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/process.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/process.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/process.cpython-39.pyc' <<< 13118 1727204064.83800: stdout chunk (state=3): >>>import 'multiprocessing.process' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211eac2100> <<< 13118 1727204064.83835: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/reduction.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/reduction.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/reduction.cpython-39.pyc' <<< 13118 1727204064.83849: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/pickle.cpython-39.pyc matches /usr/lib64/python3.9/pickle.py <<< 13118 1727204064.83890: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/pickle.cpython-39.pyc' <<< 13118 1727204064.83913: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_compat_pickle.cpython-39.pyc matches /usr/lib64/python3.9/_compat_pickle.py # code object from '/usr/lib64/python3.9/__pycache__/_compat_pickle.cpython-39.pyc' <<< 13118 1727204064.83932: stdout chunk (state=3): >>>import '_compat_pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed24a60> <<< 13118 1727204064.83943: stdout chunk (state=3): >>># extension module '_pickle' loaded from '/usr/lib64/python3.9/lib-dynload/_pickle.cpython-39-x86_64-linux-gnu.so' # extension module '_pickle' executed from '/usr/lib64/python3.9/lib-dynload/_pickle.cpython-39-x86_64-linux-gnu.so' import '_pickle' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211ed249d0> <<< 13118 1727204064.84009: stdout chunk (state=3): >>>import 'pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ecfac70> <<< 13118 1727204064.84049: stdout chunk (state=3): >>>import 'multiprocessing.reduction' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ecfac10> <<< 13118 1727204064.84059: stdout chunk (state=3): >>>import 'multiprocessing.context' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed6bbb0> import 'multiprocessing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed6bc40> <<< 13118 1727204064.84105: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/pool.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/pool.py <<< 13118 1727204064.84108: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/pool.cpython-39.pyc' <<< 13118 1727204064.84141: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/queue.cpython-39.pyc matches /usr/lib64/python3.9/queue.py # code object from '/usr/lib64/python3.9/__pycache__/queue.cpython-39.pyc' <<< 13118 1727204064.84173: stdout chunk (state=3): >>># extension module '_queue' loaded from '/usr/lib64/python3.9/lib-dynload/_queue.cpython-39-x86_64-linux-gnu.so' # extension module '_queue' executed from '/usr/lib64/python3.9/lib-dynload/_queue.cpython-39-x86_64-linux-gnu.so' import '_queue' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211ed0a310> import 'queue' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed0a9a0> <<< 13118 1727204064.84189: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/util.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/util.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/util.cpython-39.pyc' <<< 13118 1727204064.84220: stdout chunk (state=3): >>>import 'multiprocessing.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed0a940> <<< 13118 1727204064.84254: stdout chunk (state=3): >>># /usr/lib64/python3.9/multiprocessing/__pycache__/connection.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/connection.py <<< 13118 1727204064.84275: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/connection.cpython-39.pyc' <<< 13118 1727204064.84301: stdout chunk (state=3): >>># extension module '_multiprocessing' loaded from '/usr/lib64/python3.9/lib-dynload/_multiprocessing.cpython-39-x86_64-linux-gnu.so' # extension module '_multiprocessing' executed from '/usr/lib64/python3.9/lib-dynload/_multiprocessing.cpython-39-x86_64-linux-gnu.so' import '_multiprocessing' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211eb240d0> <<< 13118 1727204064.84342: stdout chunk (state=3): >>>import 'multiprocessing.connection' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7edc40> <<< 13118 1727204064.84367: stdout chunk (state=3): >>>import 'multiprocessing.pool' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed6b880> import ansible.module_utils.facts.timeout # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/timeout.py <<< 13118 1727204064.84389: stdout chunk (state=3): >>>import ansible.module_utils.facts.collector # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/collector.py # zipimport: zlib available <<< 13118 1727204064.84401: stdout chunk (state=3): >>># zipimport: zlib available import ansible.module_utils.facts.other # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/other/__init__.py # zipimport: zlib available <<< 13118 1727204064.84440: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.84495: stdout chunk (state=3): >>>import ansible.module_utils.facts.other.facter # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/other/facter.py # zipimport: zlib available <<< 13118 1727204064.84541: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.84612: stdout chunk (state=3): >>>import ansible.module_utils.facts.other.ohai # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/other/ohai.py # zipimport: zlib available <<< 13118 1727204064.84627: stdout chunk (state=3): >>># zipimport: zlib available import ansible.module_utils.facts.system # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/__init__.py # zipimport: zlib available <<< 13118 1727204064.84671: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.84686: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.apparmor # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/apparmor.py # zipimport: zlib available <<< 13118 1727204064.84730: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.84771: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.caps # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/caps.py # zipimport: zlib available <<< 13118 1727204064.84809: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.84851: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.chroot # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/chroot.py # zipimport: zlib available <<< 13118 1727204064.84904: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.84958: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.85000: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.85075: stdout chunk (state=3): >>>import ansible.module_utils.facts.utils # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/utils.py import ansible.module_utils.facts.system.cmdline # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/cmdline.py <<< 13118 1727204064.85078: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.85447: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.85804: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.distribution # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/distribution.py # zipimport: zlib available <<< 13118 1727204064.85859: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.85902: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.85951: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.85990: stdout chunk (state=3): >>>import ansible.module_utils.compat.datetime # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/compat/datetime.py import ansible.module_utils.facts.system.date_time # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/date_time.py # zipimport: zlib available <<< 13118 1727204064.86039: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.86042: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.env # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/env.py # zipimport: zlib available <<< 13118 1727204064.86084: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.86132: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.dns # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/dns.py # zipimport: zlib available <<< 13118 1727204064.86182: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.86198: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.fips # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/fips.py # zipimport: zlib available <<< 13118 1727204064.86225: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.86264: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.loadavg # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/loadavg.py # zipimport: zlib available <<< 13118 1727204064.86360: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.86395: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/glob.cpython-39.pyc matches /usr/lib64/python3.9/glob.py # code object from '/usr/lib64/python3.9/__pycache__/glob.cpython-39.pyc' <<< 13118 1727204064.86431: stdout chunk (state=3): >>>import 'glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ea15f10> <<< 13118 1727204064.86454: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/configparser.cpython-39.pyc matches /usr/lib64/python3.9/configparser.py <<< 13118 1727204064.86472: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/configparser.cpython-39.pyc' <<< 13118 1727204064.86622: stdout chunk (state=3): >>>import 'configparser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ea159d0> import ansible.module_utils.facts.system.local # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/local.py # zipimport: zlib available <<< 13118 1727204064.86686: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.86743: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.lsb # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/lsb.py # zipimport: zlib available <<< 13118 1727204064.86815: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.86894: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.pkg_mgr # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/pkg_mgr.py # zipimport: zlib available <<< 13118 1727204064.86957: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.87032: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.platform # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/platform.py # zipimport: zlib available <<< 13118 1727204064.87060: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.87113: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/ssl.cpython-39.pyc matches /usr/lib64/python3.9/ssl.py <<< 13118 1727204064.87134: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/ssl.cpython-39.pyc' <<< 13118 1727204064.87270: stdout chunk (state=3): >>># extension module '_ssl' loaded from '/usr/lib64/python3.9/lib-dynload/_ssl.cpython-39-x86_64-linux-gnu.so' # extension module '_ssl' executed from '/usr/lib64/python3.9/lib-dynload/_ssl.cpython-39-x86_64-linux-gnu.so' import '_ssl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211ea3dc10> <<< 13118 1727204064.87513: stdout chunk (state=3): >>>import 'ssl' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ea86c40> import ansible.module_utils.facts.system.python # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/python.py # zipimport: zlib available <<< 13118 1727204064.87556: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.87608: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.selinux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/selinux.py # zipimport: zlib available <<< 13118 1727204064.87686: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.87747: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.87848: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.87982: stdout chunk (state=3): >>>import ansible.module_utils.compat.version # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/compat/version.py import ansible.module_utils.facts.system.service_mgr # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/service_mgr.py # zipimport: zlib available <<< 13118 1727204064.88025: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.88060: stdout chunk (state=3): >>>import ansible.module_utils.facts.system.ssh_pub_keys # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/ssh_pub_keys.py # zipimport: zlib available <<< 13118 1727204064.88090: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.88143: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/getpass.cpython-39.pyc matches /usr/lib64/python3.9/getpass.py # code object from '/usr/lib64/python3.9/__pycache__/getpass.cpython-39.pyc' <<< 13118 1727204064.88231: stdout chunk (state=3): >>># extension module 'termios' loaded from '/usr/lib64/python3.9/lib-dynload/termios.cpython-39-x86_64-linux-gnu.so' # extension module 'termios' executed from '/usr/lib64/python3.9/lib-dynload/termios.cpython-39-x86_64-linux-gnu.so' import 'termios' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211ea885e0> import 'getpass' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ea88790> import ansible.module_utils.facts.system.user # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/user.py # zipimport: zlib available <<< 13118 1727204064.88263: stdout chunk (state=3): >>># zipimport: zlib available import ansible.module_utils.facts.hardware # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/__init__.py <<< 13118 1727204064.88268: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.88279: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.88308: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.base # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/base.py # zipimport: zlib available <<< 13118 1727204064.88443: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.88576: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.aix # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/aix.py <<< 13118 1727204064.88579: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.88651: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.88739: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.88767: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.88814: stdout chunk (state=3): >>>import ansible.module_utils.facts.sysctl # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/sysctl.py import ansible.module_utils.facts.hardware.darwin # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/darwin.py <<< 13118 1727204064.88835: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.88901: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.88937: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.89023: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.89154: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.freebsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/freebsd.py import ansible.module_utils.facts.hardware.dragonfly # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/dragonfly.py # zipimport: zlib available <<< 13118 1727204064.89256: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.89375: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.hpux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/hpux.py <<< 13118 1727204064.89387: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.89399: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.89436: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.89855: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.90283: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.linux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/linux.py import ansible.module_utils.facts.hardware.hurd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/hurd.py <<< 13118 1727204064.90287: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.90357: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.90451: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.netbsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/netbsd.py # zipimport: zlib available <<< 13118 1727204064.90540: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.90631: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.openbsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/openbsd.py # zipimport: zlib available <<< 13118 1727204064.90753: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.90915: stdout chunk (state=3): >>>import ansible.module_utils.facts.hardware.sunos # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/sunos.py # zipimport: zlib available <<< 13118 1727204064.90940: stdout chunk (state=3): >>># zipimport: zlib available import ansible.module_utils.facts.network # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/__init__.py <<< 13118 1727204064.90943: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.90960: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.90995: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.base # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/base.py # zipimport: zlib available <<< 13118 1727204064.91085: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.91158: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.91333: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.91502: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.generic_bsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/generic_bsd.py import ansible.module_utils.facts.network.aix # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/aix.py # zipimport: zlib available <<< 13118 1727204064.91542: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.91590: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.darwin # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/darwin.py <<< 13118 1727204064.91622: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.91640: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.91644: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.dragonfly # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/dragonfly.py # zipimport: zlib available <<< 13118 1727204064.91689: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.91767: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.fc_wwn # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/fc_wwn.py # zipimport: zlib available <<< 13118 1727204064.91806: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.91809: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.freebsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/freebsd.py # zipimport: zlib available <<< 13118 1727204064.91860: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.91915: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.hpux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/hpux.py <<< 13118 1727204064.91918: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.91955: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.92013: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.hurd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/hurd.py # zipimport: zlib available <<< 13118 1727204064.92233: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.92453: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.linux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/linux.py <<< 13118 1727204064.92456: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.92497: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.92554: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.iscsi # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/iscsi.py # zipimport: zlib available <<< 13118 1727204064.92589: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.92640: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.nvme # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/nvme.py <<< 13118 1727204064.92643: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.92656: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.92693: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.netbsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/netbsd.py # zipimport: zlib available <<< 13118 1727204064.92722: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.92770: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.openbsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/openbsd.py <<< 13118 1727204064.92774: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.92822: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.92924: stdout chunk (state=3): >>>import ansible.module_utils.facts.network.sunos # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/sunos.py # zipimport: zlib available <<< 13118 1727204064.92945: stdout chunk (state=3): >>># zipimport: zlib available import ansible.module_utils.facts.virtual # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/__init__.py <<< 13118 1727204064.92959: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.92974: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.93043: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.base # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/base.py # zipimport: zlib available <<< 13118 1727204064.93060: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.93063: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.93095: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.93138: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.93194: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.93280: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.sysctl # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/sysctl.py import ansible.module_utils.facts.virtual.freebsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/freebsd.py import ansible.module_utils.facts.virtual.dragonfly # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/dragonfly.py <<< 13118 1727204064.93284: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.93316: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.93366: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.hpux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/hpux.py # zipimport: zlib available <<< 13118 1727204064.93524: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.93689: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.linux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/linux.py # zipimport: zlib available <<< 13118 1727204064.93730: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.93775: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.netbsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/netbsd.py # zipimport: zlib available <<< 13118 1727204064.93823: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.93868: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.openbsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/openbsd.py <<< 13118 1727204064.93872: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.93937: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.94018: stdout chunk (state=3): >>>import ansible.module_utils.facts.virtual.sunos # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/sunos.py import ansible.module_utils.facts.default_collectors # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/default_collectors.py <<< 13118 1727204064.94022: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.94083: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.94169: stdout chunk (state=3): >>>import ansible.module_utils.facts.ansible_collector # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/ansible_collector.py import ansible.module_utils.facts.compat # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/compat.py import ansible.module_utils.facts # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/__init__.py <<< 13118 1727204064.94240: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204064.95010: stdout chunk (state=3): >>>import 'gc' # <<< 13118 1727204064.95392: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/idna.cpython-39.pyc matches /usr/lib64/python3.9/encodings/idna.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/idna.cpython-39.pyc' <<< 13118 1727204064.95429: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/stringprep.cpython-39.pyc matches /usr/lib64/python3.9/stringprep.py <<< 13118 1727204064.95432: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/stringprep.cpython-39.pyc' <<< 13118 1727204064.95452: stdout chunk (state=3): >>># extension module 'unicodedata' loaded from '/usr/lib64/python3.9/lib-dynload/unicodedata.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204064.95484: stdout chunk (state=3): >>># extension module 'unicodedata' executed from '/usr/lib64/python3.9/lib-dynload/unicodedata.cpython-39-x86_64-linux-gnu.so' import 'unicodedata' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211e828790> import 'stringprep' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211e7fad60> <<< 13118 1727204064.95542: stdout chunk (state=3): >>>import 'encodings.idna' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211e7fa6a0> <<< 13118 1727204064.95966: stdout chunk (state=3): >>> {"ansible_facts": {"ansible_apparmor": {"status": "disabled"}, "ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "root", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_local": {}, "ansible_env": {"PYTHONVERBOSE": "1", "SHELL": "/bin/bash", "PWD": "/root", "LOGNAME": "root", "XDG_SESSION_TYPE": "tty", "_": "/usr/bin/python3.9", "MOTD_SHOWN": "pam", "HOME": "/root", "LANG": "en_US.UTF-8", "LS_COLORS": "", "SSH_CONNECTION": "10.31.14.85 48676 10.31.13.78 22", "XDG_SESSION_CLASS": "user", "SELINUX_ROLE_REQUESTED": "", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "USER": "root", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "1", "XDG_SESSION_ID": "5", "XDG_RUNTIME_DIR": "/run/user/0", "SSH_CLIENT": "10.31.14.85 48676 22", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "which_declare": "declare -f", "PATH": "/root/.local/bin:/root/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "SELINUX_LEVEL_REQUESTED": "", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "SSH_TTY": "/dev/pts/0", "BASH_FUNC_which%%": "() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}"}, "ansible_python": {"version": {"major": 3, "minor": 9, "micro": 19, "releaselevel": "final", "serial": 0}, "version_info": [3, 9, 19, "final", 0], "executable": "/usr/bin/python3.9", "has_sslcontext": true, "type": "cpython"}, "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "9", "ansible_distribution_major_version": "9", "ansible_distribution_file<<< 13118 1727204064.95999: stdout chunk (state=3): >>>_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-5.14.0-511.el9.x86_64", "root": "UUID=ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": "ttyS0,115200n8"}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-5.14.0-511.el9.x86_64", "root": "UUID=ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": ["tty0", "ttyS0,115200n8"]}, "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Tuesday", "weekday_number": "2", "weeknumber": "39", "day": "24", "hour": "14", "minute": "54", "second": "24", "epoch": "1727204064", "epoch_int": "1727204064", "date": "2024-09-24", "time": "14:54:24", "iso8601_micro": "2024-09-24T18:54:24.953149Z", "iso8601": "2024-09-24T18:54:24Z", "iso8601_basic": "20240924T145424953149", "iso8601_basic_short": "20240924T145424", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAPleAC0mV69PNpLSbmzZvoLD9LsCBzX6IHRLXV1uktk0r66T6Y57EoVgflJTdo6yU0zTaJjonNzFmvC69tiRsCyywGjnvnBOvIH2vrgNGCUdVYPZbbtmQlJvol7NFFfyXQR4RSPqBKT67rYbCzbETM4j+bdDgTeDk6l7wXwz9RVvAAAAFQCuAyyjbOBDKyIW26LGcI9/nmWpHwAAAIEApIE1W6KQ7qs5kJXBdSaPoWaZUxuQhXkPWORFe7/MBn5SojDfxvJjFPo6t4QsovaCnm532Zghh1ZdB0pNm0vYcRbz3wMdfMucw/KHWt6ZEtI+sLwuMyhAVEXzmE34iXkyePtELiYzY6NyxuJ04IujI9UwD7ZnqFBHVFz529oXikIAAACBAPdUu+4Qo82CMcmrGD9vNUgtsts6GCjqBDuov8GJEALZ9ZNLlyVoNtBHLMQH9e0czLygyNGw/IDosRQkKdX4Vh4A7KXujTIOyytaN4JVJCuOBY/PeX4lreAO/UTTUJ27yT/J0Oy2Hbt+d8fZnTkZReRNPFCzvdb1nuPMG5nAyQtL", "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQCzkKXWiNuOrU77QQcZuT2T9XVh655Sh8Sv9vLWLa1uj7ceaNsB0TBiqvDFvYPENhdKceYaGAFU7sjqbmp5dlivYwPBiBWvcOgqnpBqrMG5SvP1RMiORpW6GupBLnUaMVjopPLIi0/CDlSl2eODcEnQI6BpxCCSedEKU9UrRrCFJy+6KPQXepPwKwPTd1TMzO8wpo57B5MYrjnquTNxMfgBkYsHB/V77d0tKq8qGBTkAPD8wEWLIcZOI+SyYEfCraQ95dOGAPRTFijnd7S15CugSlJ/vvcHSFXOlbgFzeNnU2jZneagkBfaOJch72opD3ebISSHCx1/kJvHN7MbksI+ljJa3Nw5LwP1XjUpT7dQMOZJDdVStXKp86K4XpWud+wMbQVVyU5QoFsCl7YTWWmSDRiPJOQI2myfizCT8i42rJ0WXm5OnqpHn1Jw4nGlcVnfgPQA/zxMldzReXdHnvriqKC9+97XgY6pj42YYP78PhOu1D2xH1AXmloNM+63VvU=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPT1h7wNcUomxtav688iXvnCnFqrHnEKf4gRaBY3w4BwbWOGxE8hq5snF9Tp+0agFeN/u980/y8BJWdWIO9Lz8I=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIPe8liWy3mh5GzCz9W616J2ArXnLOjLOZSwfmBX3Q1SI", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_fips": false, "ansible_lsb": {}, "ansible_system": "Linux", "ansible_kernel": "5.14.0-511.el9.x86_64", "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Thu Sep 19 06:52:39 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.9.19", "ansible_fqdn": "managed-node2", "ansible_hostname": "managed-node2", "ansible_nodename": "managed-node2", "ansible_domain": "", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "e28ab0e542474a869c23f7ace4640799", "ansible_service_mgr": "systemd", "ansible_pkg_mgr": "dnf", "gather_subset": ["min"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["min"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} <<< 13118 1727204064.96531: stdout chunk (state=3): >>># clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.path_hooks # clear sys.path_importer_cache # clear sys.meta_path # clear sys.__interactivehook__ # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys <<< 13118 1727204064.96546: stdout chunk (state=3): >>># cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing encodings.latin_1 # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing _locale # cleanup[2] removing _bootlocale # destroy _bootlocale # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing sre_constants # destroy sre_constants # cleanup[2] removing sre_parse # cleanup[2] removing sre_compile # cleanup[2] removing _heapq # cleanup[2] removing heapq # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib <<< 13118 1727204064.96573: stdout chunk (state=3): >>># cleanup[2] removing importlib.machinery # cleanup[2] removing collections.abc # cleanup[2] removing contextlib # cleanup[2] removing typing # destroy typing # cleanup[2] removing importlib.abc # cleanup[2] removing importlib.util # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing threading # cleanup[2] removing _bz2 # destroy _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing pwd # cleanup[2] removing grp # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing tempfile # cleanup[2] removing zipfile # destroy zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing fcntl # cleanup[2] removing locale # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing subprocess # cleanup[2] removing platform # cleanup[2] removing shlex # cleanup[2] removing token # destroy token # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 <<< 13118 1727204064.96613: stdout chunk (state=3): >>># cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # cleanup[2] removing ansible.module_utils.facts.namespace # cleanup[2] removing ansible.module_utils.compat.typing # cleanup[2] removing multiprocessing.process # cleanup[2] removing _compat_pickle # cleanup[2] removing _pickle # cleanup[2] removing pickle # cleanup[2] removing multiprocessing.reduction # cleanup[2] removing multiprocessing.context # cleanup[2] removing __mp_main__ # destroy __main__ # cleanup[2] removing multiprocessing # cleanup[2] removing _queue # cleanup[2] removing queue # cleanup[2] removing multiprocessing.util # cleanup[2] removing _multiprocessing # cleanup[2] removing multiprocessing.connection # cleanup[2] removing multiprocessing.pool # cleanup[2] removing ansible.module_utils.facts.timeout # cleanup[2] removing ansible.module_utils.facts.collector # cleanup[2] removing ansible.module_utils.facts.other # cleanup[2] removing ansible.module_utils.facts.other.facter # cleanup[2] removing ansible.module_utils.facts.other.ohai # cleanup[2] removing ansible.module_utils.facts.system # cleanup[2] removing ansible.module_utils.facts.system.apparmor # cleanup[2] removing ansible.module_utils.facts.system.caps # cleanup[2] removing ansible.module_utils.facts.system.chroot # cleanup[2] removing ansible.module_utils.facts.utils # cleanup[2] removing ansible.module_utils.facts.system.cmdline # cleanup[2] removing ansible.module_utils.facts.system.distribution # cleanup[2] removing ansible.module_utils.compat.datetime # destroy ansible.module_utils.compat.datetime # cleanup[2] removing ansible.module_utils.facts.system.date_time # cleanup[2] removing ansible.module_utils.facts.system.env # cleanup[2] removing ansible.module_utils.facts.system.dns # cleanup[2] removing ansible.module_utils.facts.system.fips <<< 13118 1727204064.96667: stdout chunk (state=3): >>># cleanup[2] removing ansible.module_utils.facts.system.loadavg # cleanup[2] removing glob # cleanup[2] removing configparser # cleanup[2] removing ansible.module_utils.facts.system.local # cleanup[2] removing ansible.module_utils.facts.system.lsb # cleanup[2] removing ansible.module_utils.facts.system.pkg_mgr # cleanup[2] removing ansible.module_utils.facts.system.platform # cleanup[2] removing _ssl # cleanup[2] removing ssl # destroy ssl # cleanup[2] removing ansible.module_utils.facts.system.python # cleanup[2] removing ansible.module_utils.facts.system.selinux # cleanup[2] removing ansible.module_utils.compat.version # destroy ansible.module_utils.compat.version # cleanup[2] removing ansible.module_utils.facts.system.service_mgr # cleanup[2] removing ansible.module_utils.facts.system.ssh_pub_keys # cleanup[2] removing termios # cleanup[2] removing getpass # cleanup[2] removing ansible.module_utils.facts.system.user # cleanup[2] removing ansible.module_utils.facts.hardware # cleanup[2] removing ansible.module_utils.facts.hardware.base # cleanup[2] removing ansible.module_utils.facts.hardware.aix # cleanup[2] removing ansible.module_utils.facts.sysctl # cleanup[2] removing ansible.module_utils.facts.hardware.darwin # cleanup[2] removing ansible.module_utils.facts.hardware.freebsd # cleanup[2] removing ansible.module_utils.facts.hardware.dragonfly # cleanup[2] removing ansible.module_utils.facts.hardware.hpux # cleanup[2] removing ansible.module_utils.facts.hardware.linux # cleanup[2] removing ansible.module_utils.facts.hardware.hurd # cleanup[2] removing ansible.module_utils.facts.hardware.netbsd # cleanup[2] removing ansible.module_utils.facts.hardware.openbsd # cleanup[2] removing ansible.module_utils.facts.hardware.sunos # cleanup[2] removing ansible.module_utils.facts.network # cleanup[2] removing ansible.module_utils.facts.network.base # cleanup[2] removing ansible.module_utils.facts.network.generic_bsd # cleanup[2] removing ansible.module_utils.facts.network.aix # cleanup[2] removing ansible.module_utils.facts.network.darwin # cleanup[2] removing ansible.module_utils.facts.network.dragonfly <<< 13118 1727204064.96683: stdout chunk (state=3): >>># cleanup[2] removing ansible.module_utils.facts.network.fc_wwn # cleanup[2] removing ansible.module_utils.facts.network.freebsd # cleanup[2] removing ansible.module_utils.facts.network.hpux # cleanup[2] removing ansible.module_utils.facts.network.hurd # cleanup[2] removing ansible.module_utils.facts.network.linux # cleanup[2] removing ansible.module_utils.facts.network.iscsi # cleanup[2] removing ansible.module_utils.facts.network.nvme # cleanup[2] removing ansible.module_utils.facts.network.netbsd # cleanup[2] removing ansible.module_utils.facts.network.openbsd # cleanup[2] removing ansible.module_utils.facts.network.sunos # cleanup[2] removing ansible.module_utils.facts.virtual # cleanup[2] removing ansible.module_utils.facts.virtual.base # cleanup[2] removing ansible.module_utils.facts.virtual.sysctl # cleanup[2] removing ansible.module_utils.facts.virtual.freebsd # cleanup[2] removing ansible.module_utils.facts.virtual.dragonfly # cleanup[2] removing ansible.module_utils.facts.virtual.hpux # cleanup[2] removing ansible.module_utils.facts.virtual.linux # cleanup[2] removing ansible.module_utils.facts.virtual.netbsd # cleanup[2] removing ansible.module_utils.facts.virtual.openbsd # cleanup[2] removing ansible.module_utils.facts.virtual.sunos # cleanup[2] removing ansible.module_utils.facts.default_collectors # cleanup[2] removing ansible.module_utils.facts.ansible_collector # cleanup[2] removing ansible.module_utils.facts.compat # cleanup[2] removing ansible.module_utils.facts # destroy ansible.module_utils.facts # destroy ansible.module_utils.facts.namespace # destroy ansible.module_utils.facts.other # destroy ansible.module_utils.facts.other.facter # destroy ansible.module_utils.facts.other.ohai # destroy ansible.module_utils.facts.system # destroy ansible.module_utils.facts.system.apparmor # destroy ansible.module_utils.facts.system.caps # destroy ansible.module_utils.facts.system.chroot # destroy ansible.module_utils.facts.system.cmdline # destroy ansible.module_utils.facts.system.distribution # destroy ansible.module_utils.facts.system.date_time # destroy ansible.module_utils.facts.system.env # destroy ansible.module_utils.facts.system.dns # destroy ansible.module_utils.facts.system.fips # destroy ansible.module_utils.facts.system.loadavg # destroy ansible.module_utils.facts.system.local # destroy ansible.module_utils.facts.system.lsb # destroy ansible.module_utils.facts.system.pkg_mgr # destroy ansible.module_utils.facts.system.platform # destroy ansible.module_utils.facts.system.python # destroy ansible.module_utils.facts.system.selinux # destroy ansible.module_utils.facts.system.service_mgr # destroy ansible.module_utils.facts.system.ssh_pub_keys # destroy ansible.module_utils.facts.system.user # destroy ansible.module_utils.facts.utils # destroy ansible.module_utils.facts.hardware # destroy ansible.module_utils.facts.hardware.base # destroy ansible.module_utils.facts.hardware.aix # destroy ansible.module_utils.facts.hardware.darwin # destroy ansible.module_utils.facts.hardware.freebsd # destroy ansible.module_utils.facts.hardware.dragonfly # destroy ansible.module_utils.facts.hardware.hpux # destroy ansible.module_utils.facts.hardware.linux # destroy ansible.module_utils.facts.hardware.hurd # destroy ansible.module_utils.facts.hardware.netbsd # destroy ansible.module_utils.facts.hardware.openbsd # destroy ansible.module_utils.facts.hardware.sunos # destroy ansible.module_utils.facts.sysctl # destroy ansible.module_utils.facts.network # destroy ansible.module_utils.facts.network.base # destroy ansible.module_utils.facts.network.generic_bsd # destroy ansible.module_utils.facts.network.aix # destroy ansible.module_utils.facts.network.darwin # destroy ansible.module_utils.facts.network.dragonfly # destroy ansible.module_utils.facts.network.fc_wwn # destroy ansible.module_utils.facts.network.freebsd # destroy ansible.module_utils.facts.network.hpux # destroy ansible.module_utils.facts.network.hurd # destroy ansible.module_utils.facts.network.linux # destroy ansible.module_utils.facts.network.iscsi # destroy ansible.module_utils.facts.network.nvme # destroy ansible.module_utils.facts.network.netbsd # destroy ansible.module_utils.facts.network.openbsd # destroy ansible.module_utils.facts.network.sunos # destroy ansible.module_utils.facts.virtual # destroy ansible.module_utils.facts.virtual.base # destroy ansible.module_utils.facts.virtual.sysctl # destroy ansible.module_utils.facts.virtual.freebsd # destroy ansible.module_utils.facts.virtual.dragonfly # destroy ansible.module_utils.facts.virtual.hpux # destroy ansible.module_utils.facts.virtual.linux # destroy ansible.module_utils.facts.virtual.netbsd # destroy ansible.module_utils.facts.virtual.openbsd # destroy ansible.module_utils.facts.virtual.sunos # destroy ansible.module_utils.facts.compat # cleanup[2] removing gc # cleanup[2] removing unicodedata # cleanup[2] removing stringprep # cleanup[2] removing encodings.idna <<< 13118 1727204064.96942: stdout chunk (state=3): >>># destroy _sitebuiltins <<< 13118 1727204064.96985: stdout chunk (state=3): >>># destroy importlib.util # destroy importlib.abc # destroy importlib.machinery <<< 13118 1727204064.97024: stdout chunk (state=3): >>># destroy zipimport # destroy _compression # destroy binascii # destroy importlib # destroy bz2 # destroy lzma <<< 13118 1727204064.97060: stdout chunk (state=3): >>># destroy __main__ # destroy locale # destroy systemd.journal # destroy systemd.daemon # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy encodings <<< 13118 1727204064.97063: stdout chunk (state=3): >>># destroy syslog # destroy uuid <<< 13118 1727204064.97095: stdout chunk (state=3): >>># destroy selinux # destroy distro # destroy logging # destroy argparse <<< 13118 1727204064.97165: stdout chunk (state=3): >>># destroy ansible.module_utils.facts.default_collectors # destroy ansible.module_utils.facts.ansible_collector # destroy multiprocessing # destroy multiprocessing.connection # destroy multiprocessing.pool <<< 13118 1727204064.97230: stdout chunk (state=3): >>># destroy pickle # destroy multiprocessing.context # destroy array # destroy _compat_pickle # destroy queue # destroy multiprocessing.process # destroy tempfile # destroy multiprocessing.util # destroy multiprocessing.reduction # destroy shlex # destroy datetime # destroy base64 <<< 13118 1727204064.97273: stdout chunk (state=3): >>># destroy ansible.module_utils.compat.selinux # destroy getpass # destroy json <<< 13118 1727204064.97278: stdout chunk (state=3): >>># destroy socket # destroy struct # destroy glob # destroy ansible.module_utils.compat.typing # destroy ansible.module_utils.facts.timeout # destroy ansible.module_utils.facts.collector <<< 13118 1727204064.97360: stdout chunk (state=3): >>># cleanup[3] wiping encodings.idna # destroy stringprep # cleanup[3] wiping unicodedata # cleanup[3] wiping gc # cleanup[3] wiping termios # cleanup[3] wiping _ssl # cleanup[3] wiping configparser # cleanup[3] wiping _multiprocessing # cleanup[3] wiping _queue # cleanup[3] wiping _pickle # cleanup[3] wiping selinux._selinux <<< 13118 1727204064.97451: stdout chunk (state=3): >>># cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # destroy configparser # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # cleanup[3] wiping tokenize # cleanup[3] wiping platform # destroy subprocess # cleanup[3] wiping selectors # cleanup[3] wiping select # cleanup[3] wiping _posixsubprocess # cleanup[3] wiping signal # cleanup[3] wiping fcntl # cleanup[3] wiping atexit # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping _blake2 # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping shutil # destroy fnmatch # cleanup[3] wiping grp # cleanup[3] wiping pwd # cleanup[3] wiping _lzma # cleanup[3] wiping threading # cleanup[3] wiping zlib # cleanup[3] wiping errno # cleanup[3] wiping weakref # cleanup[3] wiping contextlib # cleanup[3] wiping collections.abc # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap <<< 13118 1727204064.97536: stdout chunk (state=3): >>># cleanup[3] wiping _struct # cleanup[3] wiping re # destroy enum # destroy sre_compile # destroy copyreg # cleanup[3] wiping functools # cleanup[3] wiping _functools # destroy _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy heapq # destroy collections.abc # cleanup[3] wiping _collections # destroy _collections # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping itertools # cleanup[3] wiping _heapq # cleanup[3] wiping sre_parse # cleanup[3] wiping _sre # cleanup[3] wiping types # cleanup[3] wiping _locale # destroy _locale # cleanup[3] wiping os # cleanup[3] wiping os.path # destroy genericpath # cleanup[3] wiping posixpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.latin_1 # cleanup[3] wiping _signal # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix <<< 13118 1727204064.97541: stdout chunk (state=3): >>># cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins <<< 13118 1727204064.97557: stdout chunk (state=3): >>># destroy unicodedata # destroy gc # destroy termios # destroy _ssl # destroy _multiprocessing # destroy _queue # destroy _pickle # destroy systemd._daemon # destroy _socket # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy fcntl # destroy _blake2 # destroy _lzma # destroy zlib # destroy _signal <<< 13118 1727204064.97711: stdout chunk (state=3): >>># destroy platform # destroy _uuid # destroy _sre # destroy sre_parse # destroy tokenize <<< 13118 1727204064.97769: stdout chunk (state=3): >>># destroy _heapq # destroy posixpath # destroy stat # destroy ansible.module_utils.six.moves.urllib # destroy errno # destroy signal # destroy contextlib # destroy pwd # destroy grp # destroy _posixsubprocess # destroy selectors <<< 13118 1727204064.97789: stdout chunk (state=3): >>># destroy select # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy itertools # destroy operator # destroy ansible.module_utils.six.moves # destroy _operator # destroy _frozen_importlib_external # destroy _imp # destroy io # destroy marshal <<< 13118 1727204064.97815: stdout chunk (state=3): >>># destroy _frozen_importlib # clear sys.audit hooks <<< 13118 1727204064.98216: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204064.98219: stdout chunk (state=3): >>><<< 13118 1727204064.98221: stderr chunk (state=3): >>><<< 13118 1727204064.98466: _low_level_execute_command() done: rc=0, stdout=import _frozen_importlib # frozen import _imp # builtin import '_thread' # import '_warnings' # import '_weakref' # import '_io' # import 'marshal' # import 'posix' # import '_frozen_importlib_external' # # installing zipimport hook import 'time' # import 'zipimport' # # installed zipimport hook # /usr/lib64/python3.9/encodings/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/encodings/__init__.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/codecs.cpython-39.pyc matches /usr/lib64/python3.9/codecs.py # code object from '/usr/lib64/python3.9/__pycache__/codecs.cpython-39.pyc' import '_codecs' # import 'codecs' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe83dc0> # /usr/lib64/python3.9/encodings/__pycache__/aliases.cpython-39.pyc matches /usr/lib64/python3.9/encodings/aliases.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/aliases.cpython-39.pyc' import 'encodings.aliases' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe283a0> import 'encodings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe83b20> # /usr/lib64/python3.9/encodings/__pycache__/utf_8.cpython-39.pyc matches /usr/lib64/python3.9/encodings/utf_8.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/utf_8.cpython-39.pyc' import 'encodings.utf_8' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe83ac0> import '_signal' # # /usr/lib64/python3.9/encodings/__pycache__/latin_1.cpython-39.pyc matches /usr/lib64/python3.9/encodings/latin_1.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/latin_1.cpython-39.pyc' import 'encodings.latin_1' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe28490> # /usr/lib64/python3.9/__pycache__/io.cpython-39.pyc matches /usr/lib64/python3.9/io.py # code object from '/usr/lib64/python3.9/__pycache__/io.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/abc.py # code object from '/usr/lib64/python3.9/__pycache__/abc.cpython-39.pyc' import '_abc' # import 'abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe28940> import 'io' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe28670> # /usr/lib64/python3.9/__pycache__/site.cpython-39.pyc matches /usr/lib64/python3.9/site.py # code object from '/usr/lib64/python3.9/__pycache__/site.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/os.cpython-39.pyc matches /usr/lib64/python3.9/os.py # code object from '/usr/lib64/python3.9/__pycache__/os.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/stat.cpython-39.pyc matches /usr/lib64/python3.9/stat.py # code object from '/usr/lib64/python3.9/__pycache__/stat.cpython-39.pyc' import '_stat' # import 'stat' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fbcf190> # /usr/lib64/python3.9/__pycache__/_collections_abc.cpython-39.pyc matches /usr/lib64/python3.9/_collections_abc.py # code object from '/usr/lib64/python3.9/__pycache__/_collections_abc.cpython-39.pyc' import '_collections_abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fbcf220> # /usr/lib64/python3.9/__pycache__/posixpath.cpython-39.pyc matches /usr/lib64/python3.9/posixpath.py # code object from '/usr/lib64/python3.9/__pycache__/posixpath.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/genericpath.cpython-39.pyc matches /usr/lib64/python3.9/genericpath.py # code object from '/usr/lib64/python3.9/__pycache__/genericpath.cpython-39.pyc' import 'genericpath' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fbf2850> import 'posixpath' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fbcf940> import 'os' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe40880> # /usr/lib64/python3.9/__pycache__/_sitebuiltins.cpython-39.pyc matches /usr/lib64/python3.9/_sitebuiltins.py # code object from '/usr/lib64/python3.9/__pycache__/_sitebuiltins.cpython-39.pyc' import '_sitebuiltins' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fbc8d90> # /usr/lib64/python3.9/__pycache__/_bootlocale.cpython-39.pyc matches /usr/lib64/python3.9/_bootlocale.py # code object from '/usr/lib64/python3.9/__pycache__/_bootlocale.cpython-39.pyc' import '_locale' # import '_bootlocale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fbf2d90> import 'site' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fe28970> Python 3.9.19 (main, Aug 23 2024, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-2)] on linux Type "help", "copyright", "credits" or "license" for more information. # /usr/lib64/python3.9/__pycache__/base64.cpython-39.pyc matches /usr/lib64/python3.9/base64.py # code object from '/usr/lib64/python3.9/__pycache__/base64.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/re.cpython-39.pyc matches /usr/lib64/python3.9/re.py # code object from '/usr/lib64/python3.9/__pycache__/re.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/enum.cpython-39.pyc matches /usr/lib64/python3.9/enum.py # code object from '/usr/lib64/python3.9/__pycache__/enum.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/types.cpython-39.pyc matches /usr/lib64/python3.9/types.py # code object from '/usr/lib64/python3.9/__pycache__/types.cpython-39.pyc' import 'types' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb6ef10> import 'enum' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb740a0> # /usr/lib64/python3.9/__pycache__/sre_compile.cpython-39.pyc matches /usr/lib64/python3.9/sre_compile.py # code object from '/usr/lib64/python3.9/__pycache__/sre_compile.cpython-39.pyc' import '_sre' # # /usr/lib64/python3.9/__pycache__/sre_parse.cpython-39.pyc matches /usr/lib64/python3.9/sre_parse.py # code object from '/usr/lib64/python3.9/__pycache__/sre_parse.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/sre_constants.cpython-39.pyc matches /usr/lib64/python3.9/sre_constants.py # code object from '/usr/lib64/python3.9/__pycache__/sre_constants.cpython-39.pyc' import 'sre_constants' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb675b0> import 'sre_parse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb6f6a0> import 'sre_compile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb6e3d0> # /usr/lib64/python3.9/__pycache__/functools.cpython-39.pyc matches /usr/lib64/python3.9/functools.py # code object from '/usr/lib64/python3.9/__pycache__/functools.cpython-39.pyc' # /usr/lib64/python3.9/collections/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/collections/__init__.py # code object from '/usr/lib64/python3.9/collections/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/heapq.cpython-39.pyc matches /usr/lib64/python3.9/heapq.py # code object from '/usr/lib64/python3.9/__pycache__/heapq.cpython-39.pyc' # extension module '_heapq' loaded from '/usr/lib64/python3.9/lib-dynload/_heapq.cpython-39-x86_64-linux-gnu.so' # extension module '_heapq' executed from '/usr/lib64/python3.9/lib-dynload/_heapq.cpython-39-x86_64-linux-gnu.so' import '_heapq' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211fa55eb0> import 'heapq' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa559a0> import 'itertools' # # /usr/lib64/python3.9/__pycache__/keyword.cpython-39.pyc matches /usr/lib64/python3.9/keyword.py # code object from '/usr/lib64/python3.9/__pycache__/keyword.cpython-39.pyc' import 'keyword' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa55fa0> # /usr/lib64/python3.9/__pycache__/operator.cpython-39.pyc matches /usr/lib64/python3.9/operator.py # code object from '/usr/lib64/python3.9/__pycache__/operator.cpython-39.pyc' import '_operator' # import 'operator' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa55df0> # /usr/lib64/python3.9/__pycache__/reprlib.cpython-39.pyc matches /usr/lib64/python3.9/reprlib.py # code object from '/usr/lib64/python3.9/__pycache__/reprlib.cpython-39.pyc' import 'reprlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa65160> import '_collections' # import 'collections' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb49e20> import '_functools' # import 'functools' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb41700> # /usr/lib64/python3.9/__pycache__/copyreg.cpython-39.pyc matches /usr/lib64/python3.9/copyreg.py # code object from '/usr/lib64/python3.9/__pycache__/copyreg.cpython-39.pyc' import 'copyreg' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb55760> import 're' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb75eb0> # /usr/lib64/python3.9/__pycache__/struct.cpython-39.pyc matches /usr/lib64/python3.9/struct.py # code object from '/usr/lib64/python3.9/__pycache__/struct.cpython-39.pyc' # extension module '_struct' loaded from '/usr/lib64/python3.9/lib-dynload/_struct.cpython-39-x86_64-linux-gnu.so' # extension module '_struct' executed from '/usr/lib64/python3.9/lib-dynload/_struct.cpython-39-x86_64-linux-gnu.so' import '_struct' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211fa65d60> import 'struct' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb49340> # extension module 'binascii' loaded from '/usr/lib64/python3.9/lib-dynload/binascii.cpython-39-x86_64-linux-gnu.so' # extension module 'binascii' executed from '/usr/lib64/python3.9/lib-dynload/binascii.cpython-39-x86_64-linux-gnu.so' import 'binascii' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211fb55370> import 'base64' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb7ba60> # /usr/lib64/python3.9/__pycache__/runpy.cpython-39.pyc matches /usr/lib64/python3.9/runpy.py # code object from '/usr/lib64/python3.9/__pycache__/runpy.cpython-39.pyc' # /usr/lib64/python3.9/importlib/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/importlib/__init__.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/warnings.cpython-39.pyc matches /usr/lib64/python3.9/warnings.py # code object from '/usr/lib64/python3.9/__pycache__/warnings.cpython-39.pyc' import 'warnings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa65f40> import 'importlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa65e80> # /usr/lib64/python3.9/importlib/__pycache__/machinery.cpython-39.pyc matches /usr/lib64/python3.9/importlib/machinery.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/machinery.cpython-39.pyc' import 'importlib.machinery' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa65df0> # /usr/lib64/python3.9/importlib/__pycache__/util.cpython-39.pyc matches /usr/lib64/python3.9/importlib/util.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/util.cpython-39.pyc' # /usr/lib64/python3.9/importlib/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/importlib/abc.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/abc.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/typing.cpython-39.pyc matches /usr/lib64/python3.9/typing.py # code object from '/usr/lib64/python3.9/__pycache__/typing.cpython-39.pyc' # /usr/lib64/python3.9/collections/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/collections/abc.py # code object from '/usr/lib64/python3.9/collections/__pycache__/abc.cpython-39.pyc' import 'collections.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa39460> # /usr/lib64/python3.9/__pycache__/contextlib.cpython-39.pyc matches /usr/lib64/python3.9/contextlib.py # code object from '/usr/lib64/python3.9/__pycache__/contextlib.cpython-39.pyc' import 'contextlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa39550> import 'typing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa170d0> import 'importlib.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa68b20> import 'importlib.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa684c0> # /usr/lib64/python3.9/__pycache__/pkgutil.cpython-39.pyc matches /usr/lib64/python3.9/pkgutil.py # code object from '/usr/lib64/python3.9/__pycache__/pkgutil.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/weakref.cpython-39.pyc matches /usr/lib64/python3.9/weakref.py # code object from '/usr/lib64/python3.9/__pycache__/weakref.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/_weakrefset.cpython-39.pyc matches /usr/lib64/python3.9/_weakrefset.py # code object from '/usr/lib64/python3.9/__pycache__/_weakrefset.cpython-39.pyc' import '_weakrefset' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f95f2b0> import 'weakref' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa24d60> import 'pkgutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa68fa0> import 'runpy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fb7b0d0> # /usr/lib64/python3.9/__pycache__/shutil.cpython-39.pyc matches /usr/lib64/python3.9/shutil.py # code object from '/usr/lib64/python3.9/__pycache__/shutil.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/fnmatch.cpython-39.pyc matches /usr/lib64/python3.9/fnmatch.py # code object from '/usr/lib64/python3.9/__pycache__/fnmatch.cpython-39.pyc' import 'fnmatch' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f96ebe0> import 'errno' # # extension module 'zlib' loaded from '/usr/lib64/python3.9/lib-dynload/zlib.cpython-39-x86_64-linux-gnu.so' # extension module 'zlib' executed from '/usr/lib64/python3.9/lib-dynload/zlib.cpython-39-x86_64-linux-gnu.so' import 'zlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f96ef10> # /usr/lib64/python3.9/__pycache__/bz2.cpython-39.pyc matches /usr/lib64/python3.9/bz2.py # code object from '/usr/lib64/python3.9/__pycache__/bz2.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/_compression.cpython-39.pyc matches /usr/lib64/python3.9/_compression.py # code object from '/usr/lib64/python3.9/__pycache__/_compression.cpython-39.pyc' import '_compression' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f982820> # /usr/lib64/python3.9/__pycache__/threading.cpython-39.pyc matches /usr/lib64/python3.9/threading.py # code object from '/usr/lib64/python3.9/__pycache__/threading.cpython-39.pyc' import 'threading' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f982d60> # extension module '_bz2' loaded from '/usr/lib64/python3.9/lib-dynload/_bz2.cpython-39-x86_64-linux-gnu.so' # extension module '_bz2' executed from '/usr/lib64/python3.9/lib-dynload/_bz2.cpython-39-x86_64-linux-gnu.so' import '_bz2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f90f490> import 'bz2' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f96ef40> # /usr/lib64/python3.9/__pycache__/lzma.cpython-39.pyc matches /usr/lib64/python3.9/lzma.py # code object from '/usr/lib64/python3.9/__pycache__/lzma.cpython-39.pyc' # extension module '_lzma' loaded from '/usr/lib64/python3.9/lib-dynload/_lzma.cpython-39-x86_64-linux-gnu.so' # extension module '_lzma' executed from '/usr/lib64/python3.9/lib-dynload/_lzma.cpython-39-x86_64-linux-gnu.so' import '_lzma' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f91f370> import 'lzma' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f9826a0> import 'pwd' # # extension module 'grp' loaded from '/usr/lib64/python3.9/lib-dynload/grp.cpython-39-x86_64-linux-gnu.so' # extension module 'grp' executed from '/usr/lib64/python3.9/lib-dynload/grp.cpython-39-x86_64-linux-gnu.so' import 'grp' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f91f430> import 'shutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa65ac0> # /usr/lib64/python3.9/__pycache__/tempfile.cpython-39.pyc matches /usr/lib64/python3.9/tempfile.py # code object from '/usr/lib64/python3.9/__pycache__/tempfile.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/random.cpython-39.pyc matches /usr/lib64/python3.9/random.py # code object from '/usr/lib64/python3.9/__pycache__/random.cpython-39.pyc' # extension module 'math' loaded from '/usr/lib64/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so' # extension module 'math' executed from '/usr/lib64/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so' import 'math' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f93b790> # /usr/lib64/python3.9/__pycache__/bisect.cpython-39.pyc matches /usr/lib64/python3.9/bisect.py # code object from '/usr/lib64/python3.9/__pycache__/bisect.cpython-39.pyc' # extension module '_bisect' loaded from '/usr/lib64/python3.9/lib-dynload/_bisect.cpython-39-x86_64-linux-gnu.so' # extension module '_bisect' executed from '/usr/lib64/python3.9/lib-dynload/_bisect.cpython-39-x86_64-linux-gnu.so' import '_bisect' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f93ba60> import 'bisect' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f93b850> # extension module '_random' loaded from '/usr/lib64/python3.9/lib-dynload/_random.cpython-39-x86_64-linux-gnu.so' # extension module '_random' executed from '/usr/lib64/python3.9/lib-dynload/_random.cpython-39-x86_64-linux-gnu.so' import '_random' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f93b940> # /usr/lib64/python3.9/__pycache__/hashlib.cpython-39.pyc matches /usr/lib64/python3.9/hashlib.py # code object from '/usr/lib64/python3.9/__pycache__/hashlib.cpython-39.pyc' # extension module '_hashlib' loaded from '/usr/lib64/python3.9/lib-dynload/_hashlib.cpython-39-x86_64-linux-gnu.so' # extension module '_hashlib' executed from '/usr/lib64/python3.9/lib-dynload/_hashlib.cpython-39-x86_64-linux-gnu.so' import '_hashlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f93bd90> # extension module '_blake2' loaded from '/usr/lib64/python3.9/lib-dynload/_blake2.cpython-39-x86_64-linux-gnu.so' # extension module '_blake2' executed from '/usr/lib64/python3.9/lib-dynload/_blake2.cpython-39-x86_64-linux-gnu.so' import '_blake2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f9462e0> import 'hashlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f93b9d0> import 'random' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f92fb20> import 'tempfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211fa656a0> # /usr/lib64/python3.9/__pycache__/zipfile.cpython-39.pyc matches /usr/lib64/python3.9/zipfile.py # code object from '/usr/lib64/python3.9/__pycache__/zipfile.cpython-39.pyc' import 'zipfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f93bb80> # code object from '/usr/lib64/python3.9/encodings/cp437.pyc' import 'encodings.cp437' # <_frozen_importlib_external.SourcelessFileLoader object at 0x7f211f85d760> # zipimport: found 103 names in '/tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip' # zipimport: zlib available # zipimport: zlib available import ansible # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/__init__.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/__future__.cpython-39.pyc matches /usr/lib64/python3.9/__future__.py # code object from '/usr/lib64/python3.9/__pycache__/__future__.cpython-39.pyc' import '__future__' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f79b8b0> # /usr/lib64/python3.9/json/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/json/__init__.py # code object from '/usr/lib64/python3.9/json/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/json/__pycache__/decoder.cpython-39.pyc matches /usr/lib64/python3.9/json/decoder.py # code object from '/usr/lib64/python3.9/json/__pycache__/decoder.cpython-39.pyc' # /usr/lib64/python3.9/json/__pycache__/scanner.cpython-39.pyc matches /usr/lib64/python3.9/json/scanner.py # code object from '/usr/lib64/python3.9/json/__pycache__/scanner.cpython-39.pyc' # extension module '_json' loaded from '/usr/lib64/python3.9/lib-dynload/_json.cpython-39-x86_64-linux-gnu.so' # extension module '_json' executed from '/usr/lib64/python3.9/lib-dynload/_json.cpython-39-x86_64-linux-gnu.so' import '_json' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f79b160> import 'json.scanner' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f79b280> import 'json.decoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f79b5e0> # /usr/lib64/python3.9/json/__pycache__/encoder.cpython-39.pyc matches /usr/lib64/python3.9/json/encoder.py # code object from '/usr/lib64/python3.9/json/__pycache__/encoder.cpython-39.pyc' import 'json.encoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f79b4f0> import 'json' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f79be20> import 'atexit' # # extension module 'fcntl' loaded from '/usr/lib64/python3.9/lib-dynload/fcntl.cpython-39-x86_64-linux-gnu.so' # extension module 'fcntl' executed from '/usr/lib64/python3.9/lib-dynload/fcntl.cpython-39-x86_64-linux-gnu.so' import 'fcntl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f79b580> # /usr/lib64/python3.9/__pycache__/locale.cpython-39.pyc matches /usr/lib64/python3.9/locale.py # code object from '/usr/lib64/python3.9/__pycache__/locale.cpython-39.pyc' import 'locale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f79b100> # /usr/lib64/python3.9/__pycache__/platform.cpython-39.pyc matches /usr/lib64/python3.9/platform.py # code object from '/usr/lib64/python3.9/__pycache__/platform.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/subprocess.cpython-39.pyc matches /usr/lib64/python3.9/subprocess.py # code object from '/usr/lib64/python3.9/__pycache__/subprocess.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/signal.cpython-39.pyc matches /usr/lib64/python3.9/signal.py # code object from '/usr/lib64/python3.9/__pycache__/signal.cpython-39.pyc' import 'signal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f730040> # extension module '_posixsubprocess' loaded from '/usr/lib64/python3.9/lib-dynload/_posixsubprocess.cpython-39-x86_64-linux-gnu.so' # extension module '_posixsubprocess' executed from '/usr/lib64/python3.9/lib-dynload/_posixsubprocess.cpython-39-x86_64-linux-gnu.so' import '_posixsubprocess' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1493d0> # extension module 'select' loaded from '/usr/lib64/python3.9/lib-dynload/select.cpython-39-x86_64-linux-gnu.so' # extension module 'select' executed from '/usr/lib64/python3.9/lib-dynload/select.cpython-39-x86_64-linux-gnu.so' import 'select' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1490d0> # /usr/lib64/python3.9/__pycache__/selectors.cpython-39.pyc matches /usr/lib64/python3.9/selectors.py # code object from '/usr/lib64/python3.9/__pycache__/selectors.cpython-39.pyc' import 'selectors' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f149d30> import 'subprocess' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f782d90> import 'platform' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7823a0> # /usr/lib64/python3.9/__pycache__/shlex.cpython-39.pyc matches /usr/lib64/python3.9/shlex.py # code object from '/usr/lib64/python3.9/__pycache__/shlex.cpython-39.pyc' import 'shlex' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f782f40> # /usr/lib64/python3.9/__pycache__/traceback.cpython-39.pyc matches /usr/lib64/python3.9/traceback.py # code object from '/usr/lib64/python3.9/__pycache__/traceback.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/linecache.cpython-39.pyc matches /usr/lib64/python3.9/linecache.py # code object from '/usr/lib64/python3.9/__pycache__/linecache.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/tokenize.cpython-39.pyc matches /usr/lib64/python3.9/tokenize.py # code object from '/usr/lib64/python3.9/__pycache__/tokenize.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/token.cpython-39.pyc matches /usr/lib64/python3.9/token.py # code object from '/usr/lib64/python3.9/__pycache__/token.cpython-39.pyc' import 'token' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f85da90> import 'tokenize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f75adc0> import 'linecache' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f75a490> import 'traceback' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f798a90> # extension module 'syslog' loaded from '/usr/lib64/python3.9/lib-dynload/syslog.cpython-39-x86_64-linux-gnu.so' # extension module 'syslog' executed from '/usr/lib64/python3.9/lib-dynload/syslog.cpython-39-x86_64-linux-gnu.so' import 'syslog' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f75a5b0> # /usr/lib64/python3.9/site-packages/systemd/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/__init__.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/__init__.cpython-39.pyc' import 'systemd' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f75a5e0> # /usr/lib64/python3.9/site-packages/systemd/__pycache__/journal.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/journal.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/journal.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/datetime.cpython-39.pyc matches /usr/lib64/python3.9/datetime.py # code object from '/usr/lib64/python3.9/__pycache__/datetime.cpython-39.pyc' # extension module '_datetime' loaded from '/usr/lib64/python3.9/lib-dynload/_datetime.cpython-39-x86_64-linux-gnu.so' # extension module '_datetime' executed from '/usr/lib64/python3.9/lib-dynload/_datetime.cpython-39-x86_64-linux-gnu.so' import '_datetime' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1b4f70> import 'datetime' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7e52e0> # /usr/lib64/python3.9/__pycache__/uuid.cpython-39.pyc matches /usr/lib64/python3.9/uuid.py # code object from '/usr/lib64/python3.9/__pycache__/uuid.cpython-39.pyc' # extension module '_uuid' loaded from '/usr/lib64/python3.9/lib-dynload/_uuid.cpython-39-x86_64-linux-gnu.so' # extension module '_uuid' executed from '/usr/lib64/python3.9/lib-dynload/_uuid.cpython-39-x86_64-linux-gnu.so' import '_uuid' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1b17f0> import 'uuid' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7e5460> # /usr/lib64/python3.9/logging/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/logging/__init__.py # code object from '/usr/lib64/python3.9/logging/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/string.cpython-39.pyc matches /usr/lib64/python3.9/string.py # code object from '/usr/lib64/python3.9/__pycache__/string.cpython-39.pyc' import '_string' # import 'string' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7e5c40> import 'logging' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f1b1790> # extension module 'systemd._journal' loaded from '/usr/lib64/python3.9/site-packages/systemd/_journal.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._journal' executed from '/usr/lib64/python3.9/site-packages/systemd/_journal.cpython-39-x86_64-linux-gnu.so' import 'systemd._journal' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f7e5130> # extension module 'systemd._reader' loaded from '/usr/lib64/python3.9/site-packages/systemd/_reader.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._reader' executed from '/usr/lib64/python3.9/site-packages/systemd/_reader.cpython-39-x86_64-linux-gnu.so' import 'systemd._reader' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f7e5670> # extension module 'systemd.id128' loaded from '/usr/lib64/python3.9/site-packages/systemd/id128.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd.id128' executed from '/usr/lib64/python3.9/site-packages/systemd/id128.cpython-39-x86_64-linux-gnu.so' import 'systemd.id128' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f7e5730> import 'systemd.journal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7dc9a0> # /usr/lib64/python3.9/site-packages/systemd/__pycache__/daemon.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/daemon.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/daemon.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/socket.cpython-39.pyc matches /usr/lib64/python3.9/socket.py # code object from '/usr/lib64/python3.9/__pycache__/socket.cpython-39.pyc' # extension module '_socket' loaded from '/usr/lib64/python3.9/lib-dynload/_socket.cpython-39-x86_64-linux-gnu.so' # extension module '_socket' executed from '/usr/lib64/python3.9/lib-dynload/_socket.cpython-39-x86_64-linux-gnu.so' import '_socket' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1a78e0> # extension module 'array' loaded from '/usr/lib64/python3.9/lib-dynload/array.cpython-39-x86_64-linux-gnu.so' # extension module 'array' executed from '/usr/lib64/python3.9/lib-dynload/array.cpython-39-x86_64-linux-gnu.so' import 'array' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1c5c70> import 'socket' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f1b0520> # extension module 'systemd._daemon' loaded from '/usr/lib64/python3.9/site-packages/systemd/_daemon.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._daemon' executed from '/usr/lib64/python3.9/site-packages/systemd/_daemon.cpython-39-x86_64-linux-gnu.so' import 'systemd._daemon' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1a7e80> import 'systemd.daemon' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f1b0940> # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.compat # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/compat/__init__.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.text # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/text/__init__.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.six # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/six/__init__.py import 'ansible.module_utils.six.moves' # import 'ansible.module_utils.six.moves.collections_abc' # import ansible.module_utils.common.text.converters # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/text/converters.py # /usr/lib64/python3.9/ctypes/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/ctypes/__init__.py # code object from '/usr/lib64/python3.9/ctypes/__pycache__/__init__.cpython-39.pyc' # extension module '_ctypes' loaded from '/usr/lib64/python3.9/lib-dynload/_ctypes.cpython-39-x86_64-linux-gnu.so' # extension module '_ctypes' executed from '/usr/lib64/python3.9/lib-dynload/_ctypes.cpython-39-x86_64-linux-gnu.so' import '_ctypes' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1c0790> # /usr/lib64/python3.9/ctypes/__pycache__/_endian.cpython-39.pyc matches /usr/lib64/python3.9/ctypes/_endian.py # code object from '/usr/lib64/python3.9/ctypes/__pycache__/_endian.cpython-39.pyc' import 'ctypes._endian' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f1ff850> import 'ctypes' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed65fa0> import ansible.module_utils.compat.selinux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/compat/selinux.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils._text # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/_text.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/copy.cpython-39.pyc matches /usr/lib64/python3.9/copy.py # code object from '/usr/lib64/python3.9/__pycache__/copy.cpython-39.pyc' import 'copy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f762310> # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.collections # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/collections.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.warnings # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/warnings.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.errors # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/errors.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.parsing # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/parsing/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.parsing.convert_bool # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/parsing/convert_bool.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/ast.cpython-39.pyc matches /usr/lib64/python3.9/ast.py # code object from '/usr/lib64/python3.9/__pycache__/ast.cpython-39.pyc' import '_ast' # import 'ast' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7a1ca0> # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.text.formatters # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/text/formatters.py import ansible.module_utils.common.validation # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/validation.py import ansible.module_utils.common.parameters # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/parameters.py import ansible.module_utils.common.arg_spec # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/arg_spec.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.locale # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/locale.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/site-packages/selinux/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/selinux/__init__.py # code object from '/usr/lib64/python3.9/site-packages/selinux/__pycache__/__init__.cpython-39.pyc' # extension module 'selinux._selinux' loaded from '/usr/lib64/python3.9/site-packages/selinux/_selinux.cpython-39-x86_64-linux-gnu.so' # extension module 'selinux._selinux' executed from '/usr/lib64/python3.9/site-packages/selinux/_selinux.cpython-39-x86_64-linux-gnu.so' import 'selinux._selinux' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211f1e63a0> import 'selinux' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7a1bb0> import ansible.module_utils.common.file # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/file.py import ansible.module_utils.common.process # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/process.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib/python3.9/site-packages/__pycache__/distro.cpython-39.pyc matches /usr/lib/python3.9/site-packages/distro.py # code object from '/usr/lib/python3.9/site-packages/__pycache__/distro.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/argparse.cpython-39.pyc matches /usr/lib64/python3.9/argparse.py # code object from '/usr/lib64/python3.9/__pycache__/argparse.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/gettext.cpython-39.pyc matches /usr/lib64/python3.9/gettext.py # code object from '/usr/lib64/python3.9/__pycache__/gettext.cpython-39.pyc' import 'gettext' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f1c22b0> import 'argparse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f76fb80> import 'distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ebd7eb0> # destroy ansible.module_utils.distro import ansible.module_utils.distro # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/distro/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common._utils # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/_utils.py import ansible.module_utils.common.sys_info # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/common/sys_info.py import ansible.module_utils.basic # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/basic.py # zipimport: zlib available # zipimport: zlib available import ansible.modules # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/modules/__init__.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.namespace # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/namespace.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.compat.typing # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/compat/typing.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/multiprocessing/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/__init__.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/multiprocessing/__pycache__/context.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/context.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/context.cpython-39.pyc' # /usr/lib64/python3.9/multiprocessing/__pycache__/process.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/process.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/process.cpython-39.pyc' import 'multiprocessing.process' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211eac2100> # /usr/lib64/python3.9/multiprocessing/__pycache__/reduction.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/reduction.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/reduction.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/pickle.cpython-39.pyc matches /usr/lib64/python3.9/pickle.py # code object from '/usr/lib64/python3.9/__pycache__/pickle.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/_compat_pickle.cpython-39.pyc matches /usr/lib64/python3.9/_compat_pickle.py # code object from '/usr/lib64/python3.9/__pycache__/_compat_pickle.cpython-39.pyc' import '_compat_pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed24a60> # extension module '_pickle' loaded from '/usr/lib64/python3.9/lib-dynload/_pickle.cpython-39-x86_64-linux-gnu.so' # extension module '_pickle' executed from '/usr/lib64/python3.9/lib-dynload/_pickle.cpython-39-x86_64-linux-gnu.so' import '_pickle' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211ed249d0> import 'pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ecfac70> import 'multiprocessing.reduction' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ecfac10> import 'multiprocessing.context' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed6bbb0> import 'multiprocessing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed6bc40> # /usr/lib64/python3.9/multiprocessing/__pycache__/pool.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/pool.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/pool.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/queue.cpython-39.pyc matches /usr/lib64/python3.9/queue.py # code object from '/usr/lib64/python3.9/__pycache__/queue.cpython-39.pyc' # extension module '_queue' loaded from '/usr/lib64/python3.9/lib-dynload/_queue.cpython-39-x86_64-linux-gnu.so' # extension module '_queue' executed from '/usr/lib64/python3.9/lib-dynload/_queue.cpython-39-x86_64-linux-gnu.so' import '_queue' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211ed0a310> import 'queue' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed0a9a0> # /usr/lib64/python3.9/multiprocessing/__pycache__/util.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/util.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/util.cpython-39.pyc' import 'multiprocessing.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed0a940> # /usr/lib64/python3.9/multiprocessing/__pycache__/connection.cpython-39.pyc matches /usr/lib64/python3.9/multiprocessing/connection.py # code object from '/usr/lib64/python3.9/multiprocessing/__pycache__/connection.cpython-39.pyc' # extension module '_multiprocessing' loaded from '/usr/lib64/python3.9/lib-dynload/_multiprocessing.cpython-39-x86_64-linux-gnu.so' # extension module '_multiprocessing' executed from '/usr/lib64/python3.9/lib-dynload/_multiprocessing.cpython-39-x86_64-linux-gnu.so' import '_multiprocessing' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211eb240d0> import 'multiprocessing.connection' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211f7edc40> import 'multiprocessing.pool' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ed6b880> import ansible.module_utils.facts.timeout # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/timeout.py import ansible.module_utils.facts.collector # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/collector.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.other # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/other/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.other.facter # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/other/facter.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.other.ohai # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/other/ohai.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.apparmor # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/apparmor.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.caps # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/caps.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.chroot # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/chroot.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.utils # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/utils.py import ansible.module_utils.facts.system.cmdline # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/cmdline.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.distribution # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/distribution.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.compat.datetime # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/compat/datetime.py import ansible.module_utils.facts.system.date_time # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/date_time.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.env # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/env.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.dns # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/dns.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.fips # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/fips.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.loadavg # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/loadavg.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/glob.cpython-39.pyc matches /usr/lib64/python3.9/glob.py # code object from '/usr/lib64/python3.9/__pycache__/glob.cpython-39.pyc' import 'glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ea15f10> # /usr/lib64/python3.9/__pycache__/configparser.cpython-39.pyc matches /usr/lib64/python3.9/configparser.py # code object from '/usr/lib64/python3.9/__pycache__/configparser.cpython-39.pyc' import 'configparser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ea159d0> import ansible.module_utils.facts.system.local # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/local.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.lsb # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/lsb.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.pkg_mgr # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/pkg_mgr.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.platform # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/platform.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/ssl.cpython-39.pyc matches /usr/lib64/python3.9/ssl.py # code object from '/usr/lib64/python3.9/__pycache__/ssl.cpython-39.pyc' # extension module '_ssl' loaded from '/usr/lib64/python3.9/lib-dynload/_ssl.cpython-39-x86_64-linux-gnu.so' # extension module '_ssl' executed from '/usr/lib64/python3.9/lib-dynload/_ssl.cpython-39-x86_64-linux-gnu.so' import '_ssl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211ea3dc10> import 'ssl' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ea86c40> import ansible.module_utils.facts.system.python # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/python.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.selinux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/selinux.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.compat.version # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/compat/version.py import ansible.module_utils.facts.system.service_mgr # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/service_mgr.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.system.ssh_pub_keys # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/ssh_pub_keys.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/getpass.cpython-39.pyc matches /usr/lib64/python3.9/getpass.py # code object from '/usr/lib64/python3.9/__pycache__/getpass.cpython-39.pyc' # extension module 'termios' loaded from '/usr/lib64/python3.9/lib-dynload/termios.cpython-39-x86_64-linux-gnu.so' # extension module 'termios' executed from '/usr/lib64/python3.9/lib-dynload/termios.cpython-39-x86_64-linux-gnu.so' import 'termios' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211ea885e0> import 'getpass' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211ea88790> import ansible.module_utils.facts.system.user # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/system/user.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.base # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/base.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.aix # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/aix.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.sysctl # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/sysctl.py import ansible.module_utils.facts.hardware.darwin # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/darwin.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.freebsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/freebsd.py import ansible.module_utils.facts.hardware.dragonfly # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/dragonfly.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.hpux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/hpux.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.linux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/linux.py import ansible.module_utils.facts.hardware.hurd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/hurd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.netbsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/netbsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.openbsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/openbsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.hardware.sunos # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/hardware/sunos.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.base # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/base.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.generic_bsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/generic_bsd.py import ansible.module_utils.facts.network.aix # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/aix.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.darwin # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/darwin.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.dragonfly # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/dragonfly.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.fc_wwn # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/fc_wwn.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.freebsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/freebsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.hpux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/hpux.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.hurd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/hurd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.linux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/linux.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.iscsi # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/iscsi.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.nvme # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/nvme.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.netbsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/netbsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.openbsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/openbsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.network.sunos # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/network/sunos.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.base # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/base.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.sysctl # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/sysctl.py import ansible.module_utils.facts.virtual.freebsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/freebsd.py import ansible.module_utils.facts.virtual.dragonfly # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/dragonfly.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.hpux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/hpux.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.linux # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/linux.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.netbsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/netbsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.openbsd # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/openbsd.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.virtual.sunos # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/virtual/sunos.py import ansible.module_utils.facts.default_collectors # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/default_collectors.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.facts.ansible_collector # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/ansible_collector.py import ansible.module_utils.facts.compat # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/compat.py import ansible.module_utils.facts # loaded from Zip /tmp/ansible_setup_payload_icdi2pf8/ansible_setup_payload.zip/ansible/module_utils/facts/__init__.py # zipimport: zlib available import 'gc' # # /usr/lib64/python3.9/encodings/__pycache__/idna.cpython-39.pyc matches /usr/lib64/python3.9/encodings/idna.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/idna.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/stringprep.cpython-39.pyc matches /usr/lib64/python3.9/stringprep.py # code object from '/usr/lib64/python3.9/__pycache__/stringprep.cpython-39.pyc' # extension module 'unicodedata' loaded from '/usr/lib64/python3.9/lib-dynload/unicodedata.cpython-39-x86_64-linux-gnu.so' # extension module 'unicodedata' executed from '/usr/lib64/python3.9/lib-dynload/unicodedata.cpython-39-x86_64-linux-gnu.so' import 'unicodedata' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f211e828790> import 'stringprep' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211e7fad60> import 'encodings.idna' # <_frozen_importlib_external.SourceFileLoader object at 0x7f211e7fa6a0> {"ansible_facts": {"ansible_apparmor": {"status": "disabled"}, "ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "root", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_local": {}, "ansible_env": {"PYTHONVERBOSE": "1", "SHELL": "/bin/bash", "PWD": "/root", "LOGNAME": "root", "XDG_SESSION_TYPE": "tty", "_": "/usr/bin/python3.9", "MOTD_SHOWN": "pam", "HOME": "/root", "LANG": "en_US.UTF-8", "LS_COLORS": "", "SSH_CONNECTION": "10.31.14.85 48676 10.31.13.78 22", "XDG_SESSION_CLASS": "user", "SELINUX_ROLE_REQUESTED": "", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "USER": "root", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "1", "XDG_SESSION_ID": "5", "XDG_RUNTIME_DIR": "/run/user/0", "SSH_CLIENT": "10.31.14.85 48676 22", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "which_declare": "declare -f", "PATH": "/root/.local/bin:/root/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "SELINUX_LEVEL_REQUESTED": "", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "SSH_TTY": "/dev/pts/0", "BASH_FUNC_which%%": "() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}"}, "ansible_python": {"version": {"major": 3, "minor": 9, "micro": 19, "releaselevel": "final", "serial": 0}, "version_info": [3, 9, 19, "final", 0], "executable": "/usr/bin/python3.9", "has_sslcontext": true, "type": "cpython"}, "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "9", "ansible_distribution_major_version": "9", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-5.14.0-511.el9.x86_64", "root": "UUID=ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": "ttyS0,115200n8"}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-5.14.0-511.el9.x86_64", "root": "UUID=ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": ["tty0", "ttyS0,115200n8"]}, "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Tuesday", "weekday_number": "2", "weeknumber": "39", "day": "24", "hour": "14", "minute": "54", "second": "24", "epoch": "1727204064", "epoch_int": "1727204064", "date": "2024-09-24", "time": "14:54:24", "iso8601_micro": "2024-09-24T18:54:24.953149Z", "iso8601": "2024-09-24T18:54:24Z", "iso8601_basic": "20240924T145424953149", "iso8601_basic_short": "20240924T145424", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAPleAC0mV69PNpLSbmzZvoLD9LsCBzX6IHRLXV1uktk0r66T6Y57EoVgflJTdo6yU0zTaJjonNzFmvC69tiRsCyywGjnvnBOvIH2vrgNGCUdVYPZbbtmQlJvol7NFFfyXQR4RSPqBKT67rYbCzbETM4j+bdDgTeDk6l7wXwz9RVvAAAAFQCuAyyjbOBDKyIW26LGcI9/nmWpHwAAAIEApIE1W6KQ7qs5kJXBdSaPoWaZUxuQhXkPWORFe7/MBn5SojDfxvJjFPo6t4QsovaCnm532Zghh1ZdB0pNm0vYcRbz3wMdfMucw/KHWt6ZEtI+sLwuMyhAVEXzmE34iXkyePtELiYzY6NyxuJ04IujI9UwD7ZnqFBHVFz529oXikIAAACBAPdUu+4Qo82CMcmrGD9vNUgtsts6GCjqBDuov8GJEALZ9ZNLlyVoNtBHLMQH9e0czLygyNGw/IDosRQkKdX4Vh4A7KXujTIOyytaN4JVJCuOBY/PeX4lreAO/UTTUJ27yT/J0Oy2Hbt+d8fZnTkZReRNPFCzvdb1nuPMG5nAyQtL", "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQCzkKXWiNuOrU77QQcZuT2T9XVh655Sh8Sv9vLWLa1uj7ceaNsB0TBiqvDFvYPENhdKceYaGAFU7sjqbmp5dlivYwPBiBWvcOgqnpBqrMG5SvP1RMiORpW6GupBLnUaMVjopPLIi0/CDlSl2eODcEnQI6BpxCCSedEKU9UrRrCFJy+6KPQXepPwKwPTd1TMzO8wpo57B5MYrjnquTNxMfgBkYsHB/V77d0tKq8qGBTkAPD8wEWLIcZOI+SyYEfCraQ95dOGAPRTFijnd7S15CugSlJ/vvcHSFXOlbgFzeNnU2jZneagkBfaOJch72opD3ebISSHCx1/kJvHN7MbksI+ljJa3Nw5LwP1XjUpT7dQMOZJDdVStXKp86K4XpWud+wMbQVVyU5QoFsCl7YTWWmSDRiPJOQI2myfizCT8i42rJ0WXm5OnqpHn1Jw4nGlcVnfgPQA/zxMldzReXdHnvriqKC9+97XgY6pj42YYP78PhOu1D2xH1AXmloNM+63VvU=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPT1h7wNcUomxtav688iXvnCnFqrHnEKf4gRaBY3w4BwbWOGxE8hq5snF9Tp+0agFeN/u980/y8BJWdWIO9Lz8I=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIPe8liWy3mh5GzCz9W616J2ArXnLOjLOZSwfmBX3Q1SI", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_fips": false, "ansible_lsb": {}, "ansible_system": "Linux", "ansible_kernel": "5.14.0-511.el9.x86_64", "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Thu Sep 19 06:52:39 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.9.19", "ansible_fqdn": "managed-node2", "ansible_hostname": "managed-node2", "ansible_nodename": "managed-node2", "ansible_domain": "", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "e28ab0e542474a869c23f7ace4640799", "ansible_service_mgr": "systemd", "ansible_pkg_mgr": "dnf", "gather_subset": ["min"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["min"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.path_hooks # clear sys.path_importer_cache # clear sys.meta_path # clear sys.__interactivehook__ # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing encodings.latin_1 # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing _locale # cleanup[2] removing _bootlocale # destroy _bootlocale # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing sre_constants # destroy sre_constants # cleanup[2] removing sre_parse # cleanup[2] removing sre_compile # cleanup[2] removing _heapq # cleanup[2] removing heapq # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing collections.abc # cleanup[2] removing contextlib # cleanup[2] removing typing # destroy typing # cleanup[2] removing importlib.abc # cleanup[2] removing importlib.util # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing threading # cleanup[2] removing _bz2 # destroy _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing pwd # cleanup[2] removing grp # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing tempfile # cleanup[2] removing zipfile # destroy zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing fcntl # cleanup[2] removing locale # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing subprocess # cleanup[2] removing platform # cleanup[2] removing shlex # cleanup[2] removing token # destroy token # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # cleanup[2] removing ansible.module_utils.facts.namespace # cleanup[2] removing ansible.module_utils.compat.typing # cleanup[2] removing multiprocessing.process # cleanup[2] removing _compat_pickle # cleanup[2] removing _pickle # cleanup[2] removing pickle # cleanup[2] removing multiprocessing.reduction # cleanup[2] removing multiprocessing.context # cleanup[2] removing __mp_main__ # destroy __main__ # cleanup[2] removing multiprocessing # cleanup[2] removing _queue # cleanup[2] removing queue # cleanup[2] removing multiprocessing.util # cleanup[2] removing _multiprocessing # cleanup[2] removing multiprocessing.connection # cleanup[2] removing multiprocessing.pool # cleanup[2] removing ansible.module_utils.facts.timeout # cleanup[2] removing ansible.module_utils.facts.collector # cleanup[2] removing ansible.module_utils.facts.other # cleanup[2] removing ansible.module_utils.facts.other.facter # cleanup[2] removing ansible.module_utils.facts.other.ohai # cleanup[2] removing ansible.module_utils.facts.system # cleanup[2] removing ansible.module_utils.facts.system.apparmor # cleanup[2] removing ansible.module_utils.facts.system.caps # cleanup[2] removing ansible.module_utils.facts.system.chroot # cleanup[2] removing ansible.module_utils.facts.utils # cleanup[2] removing ansible.module_utils.facts.system.cmdline # cleanup[2] removing ansible.module_utils.facts.system.distribution # cleanup[2] removing ansible.module_utils.compat.datetime # destroy ansible.module_utils.compat.datetime # cleanup[2] removing ansible.module_utils.facts.system.date_time # cleanup[2] removing ansible.module_utils.facts.system.env # cleanup[2] removing ansible.module_utils.facts.system.dns # cleanup[2] removing ansible.module_utils.facts.system.fips # cleanup[2] removing ansible.module_utils.facts.system.loadavg # cleanup[2] removing glob # cleanup[2] removing configparser # cleanup[2] removing ansible.module_utils.facts.system.local # cleanup[2] removing ansible.module_utils.facts.system.lsb # cleanup[2] removing ansible.module_utils.facts.system.pkg_mgr # cleanup[2] removing ansible.module_utils.facts.system.platform # cleanup[2] removing _ssl # cleanup[2] removing ssl # destroy ssl # cleanup[2] removing ansible.module_utils.facts.system.python # cleanup[2] removing ansible.module_utils.facts.system.selinux # cleanup[2] removing ansible.module_utils.compat.version # destroy ansible.module_utils.compat.version # cleanup[2] removing ansible.module_utils.facts.system.service_mgr # cleanup[2] removing ansible.module_utils.facts.system.ssh_pub_keys # cleanup[2] removing termios # cleanup[2] removing getpass # cleanup[2] removing ansible.module_utils.facts.system.user # cleanup[2] removing ansible.module_utils.facts.hardware # cleanup[2] removing ansible.module_utils.facts.hardware.base # cleanup[2] removing ansible.module_utils.facts.hardware.aix # cleanup[2] removing ansible.module_utils.facts.sysctl # cleanup[2] removing ansible.module_utils.facts.hardware.darwin # cleanup[2] removing ansible.module_utils.facts.hardware.freebsd # cleanup[2] removing ansible.module_utils.facts.hardware.dragonfly # cleanup[2] removing ansible.module_utils.facts.hardware.hpux # cleanup[2] removing ansible.module_utils.facts.hardware.linux # cleanup[2] removing ansible.module_utils.facts.hardware.hurd # cleanup[2] removing ansible.module_utils.facts.hardware.netbsd # cleanup[2] removing ansible.module_utils.facts.hardware.openbsd # cleanup[2] removing ansible.module_utils.facts.hardware.sunos # cleanup[2] removing ansible.module_utils.facts.network # cleanup[2] removing ansible.module_utils.facts.network.base # cleanup[2] removing ansible.module_utils.facts.network.generic_bsd # cleanup[2] removing ansible.module_utils.facts.network.aix # cleanup[2] removing ansible.module_utils.facts.network.darwin # cleanup[2] removing ansible.module_utils.facts.network.dragonfly # cleanup[2] removing ansible.module_utils.facts.network.fc_wwn # cleanup[2] removing ansible.module_utils.facts.network.freebsd # cleanup[2] removing ansible.module_utils.facts.network.hpux # cleanup[2] removing ansible.module_utils.facts.network.hurd # cleanup[2] removing ansible.module_utils.facts.network.linux # cleanup[2] removing ansible.module_utils.facts.network.iscsi # cleanup[2] removing ansible.module_utils.facts.network.nvme # cleanup[2] removing ansible.module_utils.facts.network.netbsd # cleanup[2] removing ansible.module_utils.facts.network.openbsd # cleanup[2] removing ansible.module_utils.facts.network.sunos # cleanup[2] removing ansible.module_utils.facts.virtual # cleanup[2] removing ansible.module_utils.facts.virtual.base # cleanup[2] removing ansible.module_utils.facts.virtual.sysctl # cleanup[2] removing ansible.module_utils.facts.virtual.freebsd # cleanup[2] removing ansible.module_utils.facts.virtual.dragonfly # cleanup[2] removing ansible.module_utils.facts.virtual.hpux # cleanup[2] removing ansible.module_utils.facts.virtual.linux # cleanup[2] removing ansible.module_utils.facts.virtual.netbsd # cleanup[2] removing ansible.module_utils.facts.virtual.openbsd # cleanup[2] removing ansible.module_utils.facts.virtual.sunos # cleanup[2] removing ansible.module_utils.facts.default_collectors # cleanup[2] removing ansible.module_utils.facts.ansible_collector # cleanup[2] removing ansible.module_utils.facts.compat # cleanup[2] removing ansible.module_utils.facts # destroy ansible.module_utils.facts # destroy ansible.module_utils.facts.namespace # destroy ansible.module_utils.facts.other # destroy ansible.module_utils.facts.other.facter # destroy ansible.module_utils.facts.other.ohai # destroy ansible.module_utils.facts.system # destroy ansible.module_utils.facts.system.apparmor # destroy ansible.module_utils.facts.system.caps # destroy ansible.module_utils.facts.system.chroot # destroy ansible.module_utils.facts.system.cmdline # destroy ansible.module_utils.facts.system.distribution # destroy ansible.module_utils.facts.system.date_time # destroy ansible.module_utils.facts.system.env # destroy ansible.module_utils.facts.system.dns # destroy ansible.module_utils.facts.system.fips # destroy ansible.module_utils.facts.system.loadavg # destroy ansible.module_utils.facts.system.local # destroy ansible.module_utils.facts.system.lsb # destroy ansible.module_utils.facts.system.pkg_mgr # destroy ansible.module_utils.facts.system.platform # destroy ansible.module_utils.facts.system.python # destroy ansible.module_utils.facts.system.selinux # destroy ansible.module_utils.facts.system.service_mgr # destroy ansible.module_utils.facts.system.ssh_pub_keys # destroy ansible.module_utils.facts.system.user # destroy ansible.module_utils.facts.utils # destroy ansible.module_utils.facts.hardware # destroy ansible.module_utils.facts.hardware.base # destroy ansible.module_utils.facts.hardware.aix # destroy ansible.module_utils.facts.hardware.darwin # destroy ansible.module_utils.facts.hardware.freebsd # destroy ansible.module_utils.facts.hardware.dragonfly # destroy ansible.module_utils.facts.hardware.hpux # destroy ansible.module_utils.facts.hardware.linux # destroy ansible.module_utils.facts.hardware.hurd # destroy ansible.module_utils.facts.hardware.netbsd # destroy ansible.module_utils.facts.hardware.openbsd # destroy ansible.module_utils.facts.hardware.sunos # destroy ansible.module_utils.facts.sysctl # destroy ansible.module_utils.facts.network # destroy ansible.module_utils.facts.network.base # destroy ansible.module_utils.facts.network.generic_bsd # destroy ansible.module_utils.facts.network.aix # destroy ansible.module_utils.facts.network.darwin # destroy ansible.module_utils.facts.network.dragonfly # destroy ansible.module_utils.facts.network.fc_wwn # destroy ansible.module_utils.facts.network.freebsd # destroy ansible.module_utils.facts.network.hpux # destroy ansible.module_utils.facts.network.hurd # destroy ansible.module_utils.facts.network.linux # destroy ansible.module_utils.facts.network.iscsi # destroy ansible.module_utils.facts.network.nvme # destroy ansible.module_utils.facts.network.netbsd # destroy ansible.module_utils.facts.network.openbsd # destroy ansible.module_utils.facts.network.sunos # destroy ansible.module_utils.facts.virtual # destroy ansible.module_utils.facts.virtual.base # destroy ansible.module_utils.facts.virtual.sysctl # destroy ansible.module_utils.facts.virtual.freebsd # destroy ansible.module_utils.facts.virtual.dragonfly # destroy ansible.module_utils.facts.virtual.hpux # destroy ansible.module_utils.facts.virtual.linux # destroy ansible.module_utils.facts.virtual.netbsd # destroy ansible.module_utils.facts.virtual.openbsd # destroy ansible.module_utils.facts.virtual.sunos # destroy ansible.module_utils.facts.compat # cleanup[2] removing gc # cleanup[2] removing unicodedata # cleanup[2] removing stringprep # cleanup[2] removing encodings.idna # destroy _sitebuiltins # destroy importlib.util # destroy importlib.abc # destroy importlib.machinery # destroy zipimport # destroy _compression # destroy binascii # destroy importlib # destroy bz2 # destroy lzma # destroy __main__ # destroy locale # destroy systemd.journal # destroy systemd.daemon # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy encodings # destroy syslog # destroy uuid # destroy selinux # destroy distro # destroy logging # destroy argparse # destroy ansible.module_utils.facts.default_collectors # destroy ansible.module_utils.facts.ansible_collector # destroy multiprocessing # destroy multiprocessing.connection # destroy multiprocessing.pool # destroy pickle # destroy multiprocessing.context # destroy array # destroy _compat_pickle # destroy queue # destroy multiprocessing.process # destroy tempfile # destroy multiprocessing.util # destroy multiprocessing.reduction # destroy shlex # destroy datetime # destroy base64 # destroy ansible.module_utils.compat.selinux # destroy getpass # destroy json # destroy socket # destroy struct # destroy glob # destroy ansible.module_utils.compat.typing # destroy ansible.module_utils.facts.timeout # destroy ansible.module_utils.facts.collector # cleanup[3] wiping encodings.idna # destroy stringprep # cleanup[3] wiping unicodedata # cleanup[3] wiping gc # cleanup[3] wiping termios # cleanup[3] wiping _ssl # cleanup[3] wiping configparser # cleanup[3] wiping _multiprocessing # cleanup[3] wiping _queue # cleanup[3] wiping _pickle # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # destroy configparser # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # cleanup[3] wiping tokenize # cleanup[3] wiping platform # destroy subprocess # cleanup[3] wiping selectors # cleanup[3] wiping select # cleanup[3] wiping _posixsubprocess # cleanup[3] wiping signal # cleanup[3] wiping fcntl # cleanup[3] wiping atexit # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping _blake2 # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping shutil # destroy fnmatch # cleanup[3] wiping grp # cleanup[3] wiping pwd # cleanup[3] wiping _lzma # cleanup[3] wiping threading # cleanup[3] wiping zlib # cleanup[3] wiping errno # cleanup[3] wiping weakref # cleanup[3] wiping contextlib # cleanup[3] wiping collections.abc # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy enum # destroy sre_compile # destroy copyreg # cleanup[3] wiping functools # cleanup[3] wiping _functools # destroy _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy heapq # destroy collections.abc # cleanup[3] wiping _collections # destroy _collections # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping itertools # cleanup[3] wiping _heapq # cleanup[3] wiping sre_parse # cleanup[3] wiping _sre # cleanup[3] wiping types # cleanup[3] wiping _locale # destroy _locale # cleanup[3] wiping os # cleanup[3] wiping os.path # destroy genericpath # cleanup[3] wiping posixpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.latin_1 # cleanup[3] wiping _signal # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy unicodedata # destroy gc # destroy termios # destroy _ssl # destroy _multiprocessing # destroy _queue # destroy _pickle # destroy systemd._daemon # destroy _socket # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy fcntl # destroy _blake2 # destroy _lzma # destroy zlib # destroy _signal # destroy platform # destroy _uuid # destroy _sre # destroy sre_parse # destroy tokenize # destroy _heapq # destroy posixpath # destroy stat # destroy ansible.module_utils.six.moves.urllib # destroy errno # destroy signal # destroy contextlib # destroy pwd # destroy grp # destroy _posixsubprocess # destroy selectors # destroy select # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy itertools # destroy operator # destroy ansible.module_utils.six.moves # destroy _operator # destroy _frozen_importlib_external # destroy _imp # destroy io # destroy marshal # destroy _frozen_importlib # clear sys.audit hooks , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. [WARNING]: Module invocation had junk after the JSON data: # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.path_hooks # clear sys.path_importer_cache # clear sys.meta_path # clear sys.__interactivehook__ # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing encodings.latin_1 # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing _locale # cleanup[2] removing _bootlocale # destroy _bootlocale # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing sre_constants # destroy sre_constants # cleanup[2] removing sre_parse # cleanup[2] removing sre_compile # cleanup[2] removing _heapq # cleanup[2] removing heapq # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing collections.abc # cleanup[2] removing contextlib # cleanup[2] removing typing # destroy typing # cleanup[2] removing importlib.abc # cleanup[2] removing importlib.util # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing threading # cleanup[2] removing _bz2 # destroy _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing pwd # cleanup[2] removing grp # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing tempfile # cleanup[2] removing zipfile # destroy zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing fcntl # cleanup[2] removing locale # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing subprocess # cleanup[2] removing platform # cleanup[2] removing shlex # cleanup[2] removing token # destroy token # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # cleanup[2] removing ansible.module_utils.facts.namespace # cleanup[2] removing ansible.module_utils.compat.typing # cleanup[2] removing multiprocessing.process # cleanup[2] removing _compat_pickle # cleanup[2] removing _pickle # cleanup[2] removing pickle # cleanup[2] removing multiprocessing.reduction # cleanup[2] removing multiprocessing.context # cleanup[2] removing __mp_main__ # destroy __main__ # cleanup[2] removing multiprocessing # cleanup[2] removing _queue # cleanup[2] removing queue # cleanup[2] removing multiprocessing.util # cleanup[2] removing _multiprocessing # cleanup[2] removing multiprocessing.connection # cleanup[2] removing multiprocessing.pool # cleanup[2] removing ansible.module_utils.facts.timeout # cleanup[2] removing ansible.module_utils.facts.collector # cleanup[2] removing ansible.module_utils.facts.other # cleanup[2] removing ansible.module_utils.facts.other.facter # cleanup[2] removing ansible.module_utils.facts.other.ohai # cleanup[2] removing ansible.module_utils.facts.system # cleanup[2] removing ansible.module_utils.facts.system.apparmor # cleanup[2] removing ansible.module_utils.facts.system.caps # cleanup[2] removing ansible.module_utils.facts.system.chroot # cleanup[2] removing ansible.module_utils.facts.utils # cleanup[2] removing ansible.module_utils.facts.system.cmdline # cleanup[2] removing ansible.module_utils.facts.system.distribution # cleanup[2] removing ansible.module_utils.compat.datetime # destroy ansible.module_utils.compat.datetime # cleanup[2] removing ansible.module_utils.facts.system.date_time # cleanup[2] removing ansible.module_utils.facts.system.env # cleanup[2] removing ansible.module_utils.facts.system.dns # cleanup[2] removing ansible.module_utils.facts.system.fips # cleanup[2] removing ansible.module_utils.facts.system.loadavg # cleanup[2] removing glob # cleanup[2] removing configparser # cleanup[2] removing ansible.module_utils.facts.system.local # cleanup[2] removing ansible.module_utils.facts.system.lsb # cleanup[2] removing ansible.module_utils.facts.system.pkg_mgr # cleanup[2] removing ansible.module_utils.facts.system.platform # cleanup[2] removing _ssl # cleanup[2] removing ssl # destroy ssl # cleanup[2] removing ansible.module_utils.facts.system.python # cleanup[2] removing ansible.module_utils.facts.system.selinux # cleanup[2] removing ansible.module_utils.compat.version # destroy ansible.module_utils.compat.version # cleanup[2] removing ansible.module_utils.facts.system.service_mgr # cleanup[2] removing ansible.module_utils.facts.system.ssh_pub_keys # cleanup[2] removing termios # cleanup[2] removing getpass # cleanup[2] removing ansible.module_utils.facts.system.user # cleanup[2] removing ansible.module_utils.facts.hardware # cleanup[2] removing ansible.module_utils.facts.hardware.base # cleanup[2] removing ansible.module_utils.facts.hardware.aix # cleanup[2] removing ansible.module_utils.facts.sysctl # cleanup[2] removing ansible.module_utils.facts.hardware.darwin # cleanup[2] removing ansible.module_utils.facts.hardware.freebsd # cleanup[2] removing ansible.module_utils.facts.hardware.dragonfly # cleanup[2] removing ansible.module_utils.facts.hardware.hpux # cleanup[2] removing ansible.module_utils.facts.hardware.linux # cleanup[2] removing ansible.module_utils.facts.hardware.hurd # cleanup[2] removing ansible.module_utils.facts.hardware.netbsd # cleanup[2] removing ansible.module_utils.facts.hardware.openbsd # cleanup[2] removing ansible.module_utils.facts.hardware.sunos # cleanup[2] removing ansible.module_utils.facts.network # cleanup[2] removing ansible.module_utils.facts.network.base # cleanup[2] removing ansible.module_utils.facts.network.generic_bsd # cleanup[2] removing ansible.module_utils.facts.network.aix # cleanup[2] removing ansible.module_utils.facts.network.darwin # cleanup[2] removing ansible.module_utils.facts.network.dragonfly # cleanup[2] removing ansible.module_utils.facts.network.fc_wwn # cleanup[2] removing ansible.module_utils.facts.network.freebsd # cleanup[2] removing ansible.module_utils.facts.network.hpux # cleanup[2] removing ansible.module_utils.facts.network.hurd # cleanup[2] removing ansible.module_utils.facts.network.linux # cleanup[2] removing ansible.module_utils.facts.network.iscsi # cleanup[2] removing ansible.module_utils.facts.network.nvme # cleanup[2] removing ansible.module_utils.facts.network.netbsd # cleanup[2] removing ansible.module_utils.facts.network.openbsd # cleanup[2] removing ansible.module_utils.facts.network.sunos # cleanup[2] removing ansible.module_utils.facts.virtual # cleanup[2] removing ansible.module_utils.facts.virtual.base # cleanup[2] removing ansible.module_utils.facts.virtual.sysctl # cleanup[2] removing ansible.module_utils.facts.virtual.freebsd # cleanup[2] removing ansible.module_utils.facts.virtual.dragonfly # cleanup[2] removing ansible.module_utils.facts.virtual.hpux # cleanup[2] removing ansible.module_utils.facts.virtual.linux # cleanup[2] removing ansible.module_utils.facts.virtual.netbsd # cleanup[2] removing ansible.module_utils.facts.virtual.openbsd # cleanup[2] removing ansible.module_utils.facts.virtual.sunos # cleanup[2] removing ansible.module_utils.facts.default_collectors # cleanup[2] removing ansible.module_utils.facts.ansible_collector # cleanup[2] removing ansible.module_utils.facts.compat # cleanup[2] removing ansible.module_utils.facts # destroy ansible.module_utils.facts # destroy ansible.module_utils.facts.namespace # destroy ansible.module_utils.facts.other # destroy ansible.module_utils.facts.other.facter # destroy ansible.module_utils.facts.other.ohai # destroy ansible.module_utils.facts.system # destroy ansible.module_utils.facts.system.apparmor # destroy ansible.module_utils.facts.system.caps # destroy ansible.module_utils.facts.system.chroot # destroy ansible.module_utils.facts.system.cmdline # destroy ansible.module_utils.facts.system.distribution # destroy ansible.module_utils.facts.system.date_time # destroy ansible.module_utils.facts.system.env # destroy ansible.module_utils.facts.system.dns # destroy ansible.module_utils.facts.system.fips # destroy ansible.module_utils.facts.system.loadavg # destroy ansible.module_utils.facts.system.local # destroy ansible.module_utils.facts.system.lsb # destroy ansible.module_utils.facts.system.pkg_mgr # destroy ansible.module_utils.facts.system.platform # destroy ansible.module_utils.facts.system.python # destroy ansible.module_utils.facts.system.selinux # destroy ansible.module_utils.facts.system.service_mgr # destroy ansible.module_utils.facts.system.ssh_pub_keys # destroy ansible.module_utils.facts.system.user # destroy ansible.module_utils.facts.utils # destroy ansible.module_utils.facts.hardware # destroy ansible.module_utils.facts.hardware.base # destroy ansible.module_utils.facts.hardware.aix # destroy ansible.module_utils.facts.hardware.darwin # destroy ansible.module_utils.facts.hardware.freebsd # destroy ansible.module_utils.facts.hardware.dragonfly # destroy ansible.module_utils.facts.hardware.hpux # destroy ansible.module_utils.facts.hardware.linux # destroy ansible.module_utils.facts.hardware.hurd # destroy ansible.module_utils.facts.hardware.netbsd # destroy ansible.module_utils.facts.hardware.openbsd # destroy ansible.module_utils.facts.hardware.sunos # destroy ansible.module_utils.facts.sysctl # destroy ansible.module_utils.facts.network # destroy ansible.module_utils.facts.network.base # destroy ansible.module_utils.facts.network.generic_bsd # destroy ansible.module_utils.facts.network.aix # destroy ansible.module_utils.facts.network.darwin # destroy ansible.module_utils.facts.network.dragonfly # destroy ansible.module_utils.facts.network.fc_wwn # destroy ansible.module_utils.facts.network.freebsd # destroy ansible.module_utils.facts.network.hpux # destroy ansible.module_utils.facts.network.hurd # destroy ansible.module_utils.facts.network.linux # destroy ansible.module_utils.facts.network.iscsi # destroy ansible.module_utils.facts.network.nvme # destroy ansible.module_utils.facts.network.netbsd # destroy ansible.module_utils.facts.network.openbsd # destroy ansible.module_utils.facts.network.sunos # destroy ansible.module_utils.facts.virtual # destroy ansible.module_utils.facts.virtual.base # destroy ansible.module_utils.facts.virtual.sysctl # destroy ansible.module_utils.facts.virtual.freebsd # destroy ansible.module_utils.facts.virtual.dragonfly # destroy ansible.module_utils.facts.virtual.hpux # destroy ansible.module_utils.facts.virtual.linux # destroy ansible.module_utils.facts.virtual.netbsd # destroy ansible.module_utils.facts.virtual.openbsd # destroy ansible.module_utils.facts.virtual.sunos # destroy ansible.module_utils.facts.compat # cleanup[2] removing gc # cleanup[2] removing unicodedata # cleanup[2] removing stringprep # cleanup[2] removing encodings.idna # destroy _sitebuiltins # destroy importlib.util # destroy importlib.abc # destroy importlib.machinery # destroy zipimport # destroy _compression # destroy binascii # destroy importlib # destroy bz2 # destroy lzma # destroy __main__ # destroy locale # destroy systemd.journal # destroy systemd.daemon # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy encodings # destroy syslog # destroy uuid # destroy selinux # destroy distro # destroy logging # destroy argparse # destroy ansible.module_utils.facts.default_collectors # destroy ansible.module_utils.facts.ansible_collector # destroy multiprocessing # destroy multiprocessing.connection # destroy multiprocessing.pool # destroy pickle # destroy multiprocessing.context # destroy array # destroy _compat_pickle # destroy queue # destroy multiprocessing.process # destroy tempfile # destroy multiprocessing.util # destroy multiprocessing.reduction # destroy shlex # destroy datetime # destroy base64 # destroy ansible.module_utils.compat.selinux # destroy getpass # destroy json # destroy socket # destroy struct # destroy glob # destroy ansible.module_utils.compat.typing # destroy ansible.module_utils.facts.timeout # destroy ansible.module_utils.facts.collector # cleanup[3] wiping encodings.idna # destroy stringprep # cleanup[3] wiping unicodedata # cleanup[3] wiping gc # cleanup[3] wiping termios # cleanup[3] wiping _ssl # cleanup[3] wiping configparser # cleanup[3] wiping _multiprocessing # cleanup[3] wiping _queue # cleanup[3] wiping _pickle # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # destroy configparser # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # cleanup[3] wiping tokenize # cleanup[3] wiping platform # destroy subprocess # cleanup[3] wiping selectors # cleanup[3] wiping select # cleanup[3] wiping _posixsubprocess # cleanup[3] wiping signal # cleanup[3] wiping fcntl # cleanup[3] wiping atexit # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping _blake2 # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping shutil # destroy fnmatch # cleanup[3] wiping grp # cleanup[3] wiping pwd # cleanup[3] wiping _lzma # cleanup[3] wiping threading # cleanup[3] wiping zlib # cleanup[3] wiping errno # cleanup[3] wiping weakref # cleanup[3] wiping contextlib # cleanup[3] wiping collections.abc # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy enum # destroy sre_compile # destroy copyreg # cleanup[3] wiping functools # cleanup[3] wiping _functools # destroy _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy heapq # destroy collections.abc # cleanup[3] wiping _collections # destroy _collections # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping itertools # cleanup[3] wiping _heapq # cleanup[3] wiping sre_parse # cleanup[3] wiping _sre # cleanup[3] wiping types # cleanup[3] wiping _locale # destroy _locale # cleanup[3] wiping os # cleanup[3] wiping os.path # destroy genericpath # cleanup[3] wiping posixpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.latin_1 # cleanup[3] wiping _signal # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy unicodedata # destroy gc # destroy termios # destroy _ssl # destroy _multiprocessing # destroy _queue # destroy _pickle # destroy systemd._daemon # destroy _socket # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy fcntl # destroy _blake2 # destroy _lzma # destroy zlib # destroy _signal # destroy platform # destroy _uuid # destroy _sre # destroy sre_parse # destroy tokenize # destroy _heapq # destroy posixpath # destroy stat # destroy ansible.module_utils.six.moves.urllib # destroy errno # destroy signal # destroy contextlib # destroy pwd # destroy grp # destroy _posixsubprocess # destroy selectors # destroy select # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy itertools # destroy operator # destroy ansible.module_utils.six.moves # destroy _operator # destroy _frozen_importlib_external # destroy _imp # destroy io # destroy marshal # destroy _frozen_importlib # clear sys.audit hooks 13118 1727204064.99780: done with _execute_module (setup, {'gather_subset': 'min', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'setup', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204064.4576046-13581-235724493291493/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204064.99783: _low_level_execute_command(): starting 13118 1727204064.99786: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204064.4576046-13581-235724493291493/ > /dev/null 2>&1 && sleep 0' 13118 1727204065.00584: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204065.00589: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204065.00741: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.00745: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204065.00749: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.00959: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204065.00986: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204065.01146: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204065.02901: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204065.02986: stderr chunk (state=3): >>><<< 13118 1727204065.02990: stdout chunk (state=3): >>><<< 13118 1727204065.03273: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204065.03277: handler run complete 13118 1727204065.03280: variable 'ansible_facts' from source: unknown 13118 1727204065.03282: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204065.03285: variable 'ansible_facts' from source: unknown 13118 1727204065.03314: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204065.03378: attempt loop complete, returning result 13118 1727204065.03391: _execute() done 13118 1727204065.03398: dumping result to json 13118 1727204065.03414: done dumping result, returning 13118 1727204065.03427: done running TaskExecutor() for managed-node2/TASK: Gather the minimum subset of ansible_facts required by the network role test [0affcd87-79f5-56a3-0a64-0000000000dd] 13118 1727204065.03441: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000dd ok: [managed-node2] 13118 1727204065.03770: no more pending results, returning what we have 13118 1727204065.03774: results queue empty 13118 1727204065.03775: checking for any_errors_fatal 13118 1727204065.03776: done checking for any_errors_fatal 13118 1727204065.03777: checking for max_fail_percentage 13118 1727204065.03779: done checking for max_fail_percentage 13118 1727204065.03780: checking to see if all hosts have failed and the running result is not ok 13118 1727204065.03780: done checking to see if all hosts have failed 13118 1727204065.03781: getting the remaining hosts for this loop 13118 1727204065.03782: done getting the remaining hosts for this loop 13118 1727204065.03787: getting the next task for host managed-node2 13118 1727204065.03797: done getting next task for host managed-node2 13118 1727204065.03799: ^ task is: TASK: Check if system is ostree 13118 1727204065.03802: ^ state is: HOST STATE: block=2, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204065.03806: getting variables 13118 1727204065.03807: in VariableManager get_vars() 13118 1727204065.03833: Calling all_inventory to load vars for managed-node2 13118 1727204065.03836: Calling groups_inventory to load vars for managed-node2 13118 1727204065.03839: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204065.03850: Calling all_plugins_play to load vars for managed-node2 13118 1727204065.03853: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204065.03857: Calling groups_plugins_play to load vars for managed-node2 13118 1727204065.04024: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204065.04345: done with get_vars() 13118 1727204065.04356: done getting variables 13118 1727204065.04663: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000dd 13118 1727204065.04668: WORKER PROCESS EXITING TASK [Check if system is ostree] *********************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:17 Tuesday 24 September 2024 14:54:25 -0400 (0:00:00.787) 0:00:02.347 ***** 13118 1727204065.04742: entering _queue_task() for managed-node2/stat 13118 1727204065.05449: worker is 1 (out of 1 available) 13118 1727204065.05460: exiting _queue_task() for managed-node2/stat 13118 1727204065.05472: done queuing things up, now waiting for results queue to drain 13118 1727204065.05474: waiting for pending results... 13118 1727204065.05890: running TaskExecutor() for managed-node2/TASK: Check if system is ostree 13118 1727204065.05978: in run() - task 0affcd87-79f5-56a3-0a64-0000000000df 13118 1727204065.06083: variable 'ansible_search_path' from source: unknown 13118 1727204065.06087: variable 'ansible_search_path' from source: unknown 13118 1727204065.06125: calling self._execute() 13118 1727204065.06192: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.06274: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.06284: variable 'omit' from source: magic vars 13118 1727204065.07242: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204065.07855: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204065.07954: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204065.08078: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204065.08120: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204065.08240: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204065.08398: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204065.08432: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204065.08500: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204065.08789: Evaluated conditional (not __network_is_ostree is defined): True 13118 1727204065.08916: variable 'omit' from source: magic vars 13118 1727204065.08961: variable 'omit' from source: magic vars 13118 1727204065.09008: variable 'omit' from source: magic vars 13118 1727204065.09048: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204065.09155: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204065.09182: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204065.09204: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204065.09247: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204065.09377: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204065.09385: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.09393: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.09613: Set connection var ansible_timeout to 10 13118 1727204065.09632: Set connection var ansible_pipelining to False 13118 1727204065.09640: Set connection var ansible_connection to ssh 13118 1727204065.09650: Set connection var ansible_shell_executable to /bin/sh 13118 1727204065.09659: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204065.09671: Set connection var ansible_shell_type to sh 13118 1727204065.09703: variable 'ansible_shell_executable' from source: unknown 13118 1727204065.09777: variable 'ansible_connection' from source: unknown 13118 1727204065.09789: variable 'ansible_module_compression' from source: unknown 13118 1727204065.09798: variable 'ansible_shell_type' from source: unknown 13118 1727204065.09804: variable 'ansible_shell_executable' from source: unknown 13118 1727204065.09810: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.09817: variable 'ansible_pipelining' from source: unknown 13118 1727204065.09823: variable 'ansible_timeout' from source: unknown 13118 1727204065.09830: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.10076: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204065.10183: variable 'omit' from source: magic vars 13118 1727204065.10195: starting attempt loop 13118 1727204065.10222: running the handler 13118 1727204065.10240: _low_level_execute_command(): starting 13118 1727204065.10335: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204065.12131: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204065.12181: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204065.12198: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204065.12217: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204065.12262: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204065.12386: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204065.12402: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.12424: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204065.12437: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204065.12449: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204065.12462: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204065.12480: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204065.12501: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204065.12514: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204065.12525: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204065.12539: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.12730: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204065.12748: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204065.12768: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204065.12847: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204065.14480: stdout chunk (state=3): >>>/root <<< 13118 1727204065.14583: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204065.14693: stderr chunk (state=3): >>><<< 13118 1727204065.14697: stdout chunk (state=3): >>><<< 13118 1727204065.14778: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204065.14790: _low_level_execute_command(): starting 13118 1727204065.14793: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204065.1472027-13731-146327907219557 `" && echo ansible-tmp-1727204065.1472027-13731-146327907219557="` echo /root/.ansible/tmp/ansible-tmp-1727204065.1472027-13731-146327907219557 `" ) && sleep 0' 13118 1727204065.16128: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204065.16145: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204065.16160: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204065.16182: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204065.16230: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204065.16309: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204065.16324: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.16342: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204065.16354: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204065.16367: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204065.16379: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204065.16393: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204065.16411: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204065.16426: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204065.16436: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204065.16449: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.16524: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204065.16660: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204065.16680: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204065.16754: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204065.18654: stdout chunk (state=3): >>>ansible-tmp-1727204065.1472027-13731-146327907219557=/root/.ansible/tmp/ansible-tmp-1727204065.1472027-13731-146327907219557 <<< 13118 1727204065.18786: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204065.18881: stderr chunk (state=3): >>><<< 13118 1727204065.18885: stdout chunk (state=3): >>><<< 13118 1727204065.19074: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204065.1472027-13731-146327907219557=/root/.ansible/tmp/ansible-tmp-1727204065.1472027-13731-146327907219557 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204065.19078: variable 'ansible_module_compression' from source: unknown 13118 1727204065.19081: ANSIBALLZ: Using lock for stat 13118 1727204065.19083: ANSIBALLZ: Acquiring lock 13118 1727204065.19085: ANSIBALLZ: Lock acquired: 140051944307744 13118 1727204065.19087: ANSIBALLZ: Creating module 13118 1727204065.42775: ANSIBALLZ: Writing module into payload 13118 1727204065.42913: ANSIBALLZ: Writing module 13118 1727204065.42937: ANSIBALLZ: Renaming module 13118 1727204065.42941: ANSIBALLZ: Done creating module 13118 1727204065.42959: variable 'ansible_facts' from source: unknown 13118 1727204065.43040: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204065.1472027-13731-146327907219557/AnsiballZ_stat.py 13118 1727204065.43217: Sending initial data 13118 1727204065.43220: Sent initial data (153 bytes) 13118 1727204065.44298: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204065.44308: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204065.44321: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204065.44346: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204065.44387: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204065.44394: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204065.44404: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.44419: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204065.44426: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204065.44437: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204065.44452: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204065.44461: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204065.44478: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204065.44484: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204065.44492: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204065.44501: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.44583: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204065.44600: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204065.44608: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204065.44682: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 4 <<< 13118 1727204065.46696: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204065.46706: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpoyas0mp7 /root/.ansible/tmp/ansible-tmp-1727204065.1472027-13731-146327907219557/AnsiballZ_stat.py <<< 13118 1727204065.46726: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204065.48046: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204065.48227: stderr chunk (state=3): >>><<< 13118 1727204065.48231: stdout chunk (state=3): >>><<< 13118 1727204065.48233: done transferring module to remote 13118 1727204065.48236: _low_level_execute_command(): starting 13118 1727204065.48238: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204065.1472027-13731-146327907219557/ /root/.ansible/tmp/ansible-tmp-1727204065.1472027-13731-146327907219557/AnsiballZ_stat.py && sleep 0' 13118 1727204065.49041: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204065.49045: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204065.49087: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.49090: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204065.49093: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.49155: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204065.50088: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204065.50093: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204065.50144: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204065.51874: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204065.51957: stderr chunk (state=3): >>><<< 13118 1727204065.51961: stdout chunk (state=3): >>><<< 13118 1727204065.52076: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204065.52081: _low_level_execute_command(): starting 13118 1727204065.52083: _low_level_execute_command(): executing: /bin/sh -c 'PYTHONVERBOSE=1 /usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204065.1472027-13731-146327907219557/AnsiballZ_stat.py && sleep 0' 13118 1727204065.53166: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204065.53173: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204065.53201: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204065.53205: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.53271: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204065.54092: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204065.54099: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204065.54156: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204065.56101: stdout chunk (state=3): >>>import _frozen_importlib # frozen import _imp # builtin <<< 13118 1727204065.56106: stdout chunk (state=3): >>>import '_thread' # import '_warnings' # import '_weakref' # <<< 13118 1727204065.56166: stdout chunk (state=3): >>>import '_io' # import 'marshal' # <<< 13118 1727204065.56197: stdout chunk (state=3): >>>import 'posix' # <<< 13118 1727204065.56230: stdout chunk (state=3): >>>import '_frozen_importlib_external' # # installing zipimport hook <<< 13118 1727204065.56267: stdout chunk (state=3): >>>import 'time' # <<< 13118 1727204065.56281: stdout chunk (state=3): >>>import 'zipimport' # # installed zipimport hook <<< 13118 1727204065.56335: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/encodings/__init__.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204065.56360: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/codecs.cpython-39.pyc matches /usr/lib64/python3.9/codecs.py <<< 13118 1727204065.56389: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/codecs.cpython-39.pyc' <<< 13118 1727204065.56392: stdout chunk (state=3): >>>import '_codecs' # <<< 13118 1727204065.56417: stdout chunk (state=3): >>>import 'codecs' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c11b3dc0> <<< 13118 1727204065.56475: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/aliases.cpython-39.pyc matches /usr/lib64/python3.9/encodings/aliases.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/aliases.cpython-39.pyc' import 'encodings.aliases' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c11583a0> import 'encodings' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c11b3b20> <<< 13118 1727204065.56506: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/utf_8.cpython-39.pyc matches /usr/lib64/python3.9/encodings/utf_8.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/utf_8.cpython-39.pyc' <<< 13118 1727204065.56525: stdout chunk (state=3): >>>import 'encodings.utf_8' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c11b3ac0> <<< 13118 1727204065.56542: stdout chunk (state=3): >>>import '_signal' # <<< 13118 1727204065.56575: stdout chunk (state=3): >>># /usr/lib64/python3.9/encodings/__pycache__/latin_1.cpython-39.pyc matches /usr/lib64/python3.9/encodings/latin_1.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/latin_1.cpython-39.pyc' <<< 13118 1727204065.56591: stdout chunk (state=3): >>>import 'encodings.latin_1' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1158490> <<< 13118 1727204065.56614: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/io.cpython-39.pyc matches /usr/lib64/python3.9/io.py # code object from '/usr/lib64/python3.9/__pycache__/io.cpython-39.pyc' <<< 13118 1727204065.56639: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/abc.py # code object from '/usr/lib64/python3.9/__pycache__/abc.cpython-39.pyc' <<< 13118 1727204065.56662: stdout chunk (state=3): >>>import '_abc' # <<< 13118 1727204065.56677: stdout chunk (state=3): >>>import 'abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1158940> <<< 13118 1727204065.56699: stdout chunk (state=3): >>>import 'io' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1158670> <<< 13118 1727204065.56731: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/site.cpython-39.pyc matches /usr/lib64/python3.9/site.py <<< 13118 1727204065.56742: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/site.cpython-39.pyc' <<< 13118 1727204065.56761: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/os.cpython-39.pyc matches /usr/lib64/python3.9/os.py <<< 13118 1727204065.56782: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/os.cpython-39.pyc' <<< 13118 1727204065.56810: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/stat.cpython-39.pyc matches /usr/lib64/python3.9/stat.py <<< 13118 1727204065.56821: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/stat.cpython-39.pyc' <<< 13118 1727204065.56846: stdout chunk (state=3): >>>import '_stat' # <<< 13118 1727204065.56849: stdout chunk (state=3): >>>import 'stat' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c110f190> <<< 13118 1727204065.56874: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_collections_abc.cpython-39.pyc matches /usr/lib64/python3.9/_collections_abc.py <<< 13118 1727204065.56888: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/_collections_abc.cpython-39.pyc' <<< 13118 1727204065.56962: stdout chunk (state=3): >>>import '_collections_abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c110f220> <<< 13118 1727204065.56991: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/posixpath.cpython-39.pyc matches /usr/lib64/python3.9/posixpath.py <<< 13118 1727204065.56994: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/posixpath.cpython-39.pyc' <<< 13118 1727204065.57033: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/genericpath.cpython-39.pyc matches /usr/lib64/python3.9/genericpath.py # code object from '/usr/lib64/python3.9/__pycache__/genericpath.cpython-39.pyc' <<< 13118 1727204065.57036: stdout chunk (state=3): >>>import 'genericpath' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1132850> import 'posixpath' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c110f940> <<< 13118 1727204065.57069: stdout chunk (state=3): >>>import 'os' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1170880> <<< 13118 1727204065.57095: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_sitebuiltins.cpython-39.pyc matches /usr/lib64/python3.9/_sitebuiltins.py # code object from '/usr/lib64/python3.9/__pycache__/_sitebuiltins.cpython-39.pyc' <<< 13118 1727204065.57098: stdout chunk (state=3): >>>import '_sitebuiltins' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1108d90> <<< 13118 1727204065.57168: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_bootlocale.cpython-39.pyc matches /usr/lib64/python3.9/_bootlocale.py # code object from '/usr/lib64/python3.9/__pycache__/_bootlocale.cpython-39.pyc' <<< 13118 1727204065.57172: stdout chunk (state=3): >>>import '_locale' # <<< 13118 1727204065.57174: stdout chunk (state=3): >>>import '_bootlocale' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1132d90> <<< 13118 1727204065.57231: stdout chunk (state=3): >>>import 'site' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1158970> <<< 13118 1727204065.57256: stdout chunk (state=3): >>>Python 3.9.19 (main, Aug 23 2024, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-2)] on linux Type "help", "copyright", "credits" or "license" for more information. <<< 13118 1727204065.57455: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/base64.cpython-39.pyc matches /usr/lib64/python3.9/base64.py <<< 13118 1727204065.57477: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/base64.cpython-39.pyc' <<< 13118 1727204065.57498: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/re.cpython-39.pyc matches /usr/lib64/python3.9/re.py <<< 13118 1727204065.57501: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/re.cpython-39.pyc' <<< 13118 1727204065.57515: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/enum.cpython-39.pyc matches /usr/lib64/python3.9/enum.py <<< 13118 1727204065.57538: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/enum.cpython-39.pyc' <<< 13118 1727204065.57556: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/types.cpython-39.pyc matches /usr/lib64/python3.9/types.py <<< 13118 1727204065.57585: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/types.cpython-39.pyc' <<< 13118 1727204065.57588: stdout chunk (state=3): >>>import 'types' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10adf10> <<< 13118 1727204065.57633: stdout chunk (state=3): >>>import 'enum' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10b40a0> <<< 13118 1727204065.57647: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/sre_compile.cpython-39.pyc matches /usr/lib64/python3.9/sre_compile.py <<< 13118 1727204065.57662: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/sre_compile.cpython-39.pyc' <<< 13118 1727204065.57703: stdout chunk (state=3): >>>import '_sre' # <<< 13118 1727204065.57706: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/sre_parse.cpython-39.pyc matches /usr/lib64/python3.9/sre_parse.py <<< 13118 1727204065.57716: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/sre_parse.cpython-39.pyc' <<< 13118 1727204065.57742: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/sre_constants.cpython-39.pyc matches /usr/lib64/python3.9/sre_constants.py # code object from '/usr/lib64/python3.9/__pycache__/sre_constants.cpython-39.pyc' <<< 13118 1727204065.57769: stdout chunk (state=3): >>>import 'sre_constants' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10a75b0> <<< 13118 1727204065.57786: stdout chunk (state=3): >>>import 'sre_parse' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10ae6a0> <<< 13118 1727204065.57816: stdout chunk (state=3): >>>import 'sre_compile' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10ad3d0> <<< 13118 1727204065.57819: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/functools.cpython-39.pyc matches /usr/lib64/python3.9/functools.py <<< 13118 1727204065.57886: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/functools.cpython-39.pyc' <<< 13118 1727204065.57910: stdout chunk (state=3): >>># /usr/lib64/python3.9/collections/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/collections/__init__.py <<< 13118 1727204065.57937: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/collections/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204065.57953: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/heapq.cpython-39.pyc matches /usr/lib64/python3.9/heapq.py <<< 13118 1727204065.57967: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/heapq.cpython-39.pyc' <<< 13118 1727204065.58002: stdout chunk (state=3): >>># extension module '_heapq' loaded from '/usr/lib64/python3.9/lib-dynload/_heapq.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204065.58008: stdout chunk (state=3): >>># extension module '_heapq' executed from '/usr/lib64/python3.9/lib-dynload/_heapq.cpython-39-x86_64-linux-gnu.so' import '_heapq' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c1031eb0> import 'heapq' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10319a0> <<< 13118 1727204065.58028: stdout chunk (state=3): >>>import 'itertools' # <<< 13118 1727204065.58051: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/keyword.cpython-39.pyc matches /usr/lib64/python3.9/keyword.py <<< 13118 1727204065.58054: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/keyword.cpython-39.pyc' import 'keyword' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1031fa0> <<< 13118 1727204065.58083: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/operator.cpython-39.pyc matches /usr/lib64/python3.9/operator.py <<< 13118 1727204065.58086: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/operator.cpython-39.pyc' <<< 13118 1727204065.58111: stdout chunk (state=3): >>>import '_operator' # import 'operator' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1031df0> <<< 13118 1727204065.58156: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/reprlib.cpython-39.pyc matches /usr/lib64/python3.9/reprlib.py # code object from '/usr/lib64/python3.9/__pycache__/reprlib.cpython-39.pyc' <<< 13118 1727204065.58159: stdout chunk (state=3): >>>import 'reprlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1041160> <<< 13118 1727204065.58172: stdout chunk (state=3): >>>import '_collections' # <<< 13118 1727204065.58213: stdout chunk (state=3): >>>import 'collections' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1089e20> import '_functools' # <<< 13118 1727204065.58247: stdout chunk (state=3): >>>import 'functools' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1081700> <<< 13118 1727204065.58308: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/copyreg.cpython-39.pyc matches /usr/lib64/python3.9/copyreg.py <<< 13118 1727204065.58311: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/copyreg.cpython-39.pyc' import 'copyreg' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1095760> <<< 13118 1727204065.58314: stdout chunk (state=3): >>>import 're' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10b5eb0> <<< 13118 1727204065.58335: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/struct.cpython-39.pyc matches /usr/lib64/python3.9/struct.py # code object from '/usr/lib64/python3.9/__pycache__/struct.cpython-39.pyc' <<< 13118 1727204065.58385: stdout chunk (state=3): >>># extension module '_struct' loaded from '/usr/lib64/python3.9/lib-dynload/_struct.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204065.58388: stdout chunk (state=3): >>># extension module '_struct' executed from '/usr/lib64/python3.9/lib-dynload/_struct.cpython-39-x86_64-linux-gnu.so' import '_struct' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c1041d60> import 'struct' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1089340> <<< 13118 1727204065.58427: stdout chunk (state=3): >>># extension module 'binascii' loaded from '/usr/lib64/python3.9/lib-dynload/binascii.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204065.58430: stdout chunk (state=3): >>># extension module 'binascii' executed from '/usr/lib64/python3.9/lib-dynload/binascii.cpython-39-x86_64-linux-gnu.so' import 'binascii' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c1095370> <<< 13118 1727204065.58433: stdout chunk (state=3): >>>import 'base64' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10bba60> <<< 13118 1727204065.58461: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/runpy.cpython-39.pyc matches /usr/lib64/python3.9/runpy.py <<< 13118 1727204065.58466: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/runpy.cpython-39.pyc' <<< 13118 1727204065.58500: stdout chunk (state=3): >>># /usr/lib64/python3.9/importlib/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/importlib/__init__.py <<< 13118 1727204065.58505: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/importlib/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204065.58520: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/warnings.cpython-39.pyc matches /usr/lib64/python3.9/warnings.py # code object from '/usr/lib64/python3.9/__pycache__/warnings.cpython-39.pyc' <<< 13118 1727204065.58541: stdout chunk (state=3): >>>import 'warnings' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1041f40> <<< 13118 1727204065.58544: stdout chunk (state=3): >>>import 'importlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1041e80> <<< 13118 1727204065.58569: stdout chunk (state=3): >>># /usr/lib64/python3.9/importlib/__pycache__/machinery.cpython-39.pyc matches /usr/lib64/python3.9/importlib/machinery.py <<< 13118 1727204065.58581: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/importlib/__pycache__/machinery.cpython-39.pyc' import 'importlib.machinery' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1041df0> <<< 13118 1727204065.58607: stdout chunk (state=3): >>># /usr/lib64/python3.9/importlib/__pycache__/util.cpython-39.pyc matches /usr/lib64/python3.9/importlib/util.py <<< 13118 1727204065.58624: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/importlib/__pycache__/util.cpython-39.pyc' <<< 13118 1727204065.58629: stdout chunk (state=3): >>># /usr/lib64/python3.9/importlib/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/importlib/abc.py <<< 13118 1727204065.58641: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/importlib/__pycache__/abc.cpython-39.pyc' <<< 13118 1727204065.58656: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/typing.cpython-39.pyc matches /usr/lib64/python3.9/typing.py <<< 13118 1727204065.58711: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/typing.cpython-39.pyc' <<< 13118 1727204065.58744: stdout chunk (state=3): >>># /usr/lib64/python3.9/collections/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/collections/abc.py <<< 13118 1727204065.58747: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/collections/__pycache__/abc.cpython-39.pyc' import 'collections.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0dfa460> <<< 13118 1727204065.58765: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/contextlib.cpython-39.pyc matches /usr/lib64/python3.9/contextlib.py <<< 13118 1727204065.58781: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/contextlib.cpython-39.pyc' <<< 13118 1727204065.58810: stdout chunk (state=3): >>>import 'contextlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0dfa550> <<< 13118 1727204065.58941: stdout chunk (state=3): >>>import 'typing' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0dd80d0> <<< 13118 1727204065.58971: stdout chunk (state=3): >>>import 'importlib.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1044b20> <<< 13118 1727204065.58984: stdout chunk (state=3): >>>import 'importlib.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10444c0> <<< 13118 1727204065.58999: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/pkgutil.cpython-39.pyc matches /usr/lib64/python3.9/pkgutil.py <<< 13118 1727204065.59015: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/pkgutil.cpython-39.pyc' <<< 13118 1727204065.59058: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/weakref.cpython-39.pyc matches /usr/lib64/python3.9/weakref.py <<< 13118 1727204065.59083: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/weakref.cpython-39.pyc' <<< 13118 1727204065.59087: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_weakrefset.cpython-39.pyc matches /usr/lib64/python3.9/_weakrefset.py <<< 13118 1727204065.59099: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/_weakrefset.cpython-39.pyc' import '_weakrefset' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0d142b0> <<< 13118 1727204065.59141: stdout chunk (state=3): >>>import 'weakref' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0de5d60> <<< 13118 1727204065.59193: stdout chunk (state=3): >>>import 'pkgutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1044fa0> <<< 13118 1727204065.59196: stdout chunk (state=3): >>>import 'runpy' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10bb0d0> <<< 13118 1727204065.59211: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/shutil.cpython-39.pyc matches /usr/lib64/python3.9/shutil.py <<< 13118 1727204065.59238: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/shutil.cpython-39.pyc' <<< 13118 1727204065.59266: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/fnmatch.cpython-39.pyc matches /usr/lib64/python3.9/fnmatch.py <<< 13118 1727204065.59269: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/fnmatch.cpython-39.pyc' <<< 13118 1727204065.59287: stdout chunk (state=3): >>>import 'fnmatch' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0d24be0> import 'errno' # <<< 13118 1727204065.59328: stdout chunk (state=3): >>># extension module 'zlib' loaded from '/usr/lib64/python3.9/lib-dynload/zlib.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204065.59331: stdout chunk (state=3): >>># extension module 'zlib' executed from '/usr/lib64/python3.9/lib-dynload/zlib.cpython-39-x86_64-linux-gnu.so' import 'zlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0d24f10> <<< 13118 1727204065.59349: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/bz2.cpython-39.pyc matches /usr/lib64/python3.9/bz2.py <<< 13118 1727204065.59360: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/bz2.cpython-39.pyc' <<< 13118 1727204065.59390: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/_compression.cpython-39.pyc matches /usr/lib64/python3.9/_compression.py <<< 13118 1727204065.59401: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/_compression.cpython-39.pyc' import '_compression' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0d37820> <<< 13118 1727204065.59423: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/threading.cpython-39.pyc matches /usr/lib64/python3.9/threading.py <<< 13118 1727204065.59452: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/threading.cpython-39.pyc' <<< 13118 1727204065.59483: stdout chunk (state=3): >>>import 'threading' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0d37d60> <<< 13118 1727204065.59516: stdout chunk (state=3): >>># extension module '_bz2' loaded from '/usr/lib64/python3.9/lib-dynload/_bz2.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204065.59531: stdout chunk (state=3): >>># extension module '_bz2' executed from '/usr/lib64/python3.9/lib-dynload/_bz2.cpython-39-x86_64-linux-gnu.so' import '_bz2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cc5490> <<< 13118 1727204065.59541: stdout chunk (state=3): >>>import 'bz2' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0d24f40> <<< 13118 1727204065.59554: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/lzma.cpython-39.pyc matches /usr/lib64/python3.9/lzma.py <<< 13118 1727204065.59568: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/lzma.cpython-39.pyc' <<< 13118 1727204065.59615: stdout chunk (state=3): >>># extension module '_lzma' loaded from '/usr/lib64/python3.9/lib-dynload/_lzma.cpython-39-x86_64-linux-gnu.so' # extension module '_lzma' executed from '/usr/lib64/python3.9/lib-dynload/_lzma.cpython-39-x86_64-linux-gnu.so' import '_lzma' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cd5370> <<< 13118 1727204065.59631: stdout chunk (state=3): >>>import 'lzma' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0d376a0> import 'pwd' # <<< 13118 1727204065.59659: stdout chunk (state=3): >>># extension module 'grp' loaded from '/usr/lib64/python3.9/lib-dynload/grp.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204065.59672: stdout chunk (state=3): >>># extension module 'grp' executed from '/usr/lib64/python3.9/lib-dynload/grp.cpython-39-x86_64-linux-gnu.so' import 'grp' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cd5430> <<< 13118 1727204065.59715: stdout chunk (state=3): >>>import 'shutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1041ac0> <<< 13118 1727204065.59732: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/tempfile.cpython-39.pyc matches /usr/lib64/python3.9/tempfile.py <<< 13118 1727204065.59746: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/tempfile.cpython-39.pyc' <<< 13118 1727204065.59766: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/random.cpython-39.pyc matches /usr/lib64/python3.9/random.py <<< 13118 1727204065.59781: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/random.cpython-39.pyc' <<< 13118 1727204065.59823: stdout chunk (state=3): >>># extension module 'math' loaded from '/usr/lib64/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so' # extension module 'math' executed from '/usr/lib64/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so' import 'math' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cf1790> <<< 13118 1727204065.59842: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/bisect.cpython-39.pyc matches /usr/lib64/python3.9/bisect.py # code object from '/usr/lib64/python3.9/__pycache__/bisect.cpython-39.pyc' <<< 13118 1727204065.59871: stdout chunk (state=3): >>># extension module '_bisect' loaded from '/usr/lib64/python3.9/lib-dynload/_bisect.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204065.59883: stdout chunk (state=3): >>># extension module '_bisect' executed from '/usr/lib64/python3.9/lib-dynload/_bisect.cpython-39-x86_64-linux-gnu.so' import '_bisect' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cf1a60> import 'bisect' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0cf1850> <<< 13118 1727204065.59906: stdout chunk (state=3): >>># extension module '_random' loaded from '/usr/lib64/python3.9/lib-dynload/_random.cpython-39-x86_64-linux-gnu.so' # extension module '_random' executed from '/usr/lib64/python3.9/lib-dynload/_random.cpython-39-x86_64-linux-gnu.so' import '_random' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cf1940> <<< 13118 1727204065.59944: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/hashlib.cpython-39.pyc matches /usr/lib64/python3.9/hashlib.py # code object from '/usr/lib64/python3.9/__pycache__/hashlib.cpython-39.pyc' <<< 13118 1727204065.60148: stdout chunk (state=3): >>># extension module '_hashlib' loaded from '/usr/lib64/python3.9/lib-dynload/_hashlib.cpython-39-x86_64-linux-gnu.so' # extension module '_hashlib' executed from '/usr/lib64/python3.9/lib-dynload/_hashlib.cpython-39-x86_64-linux-gnu.so' import '_hashlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cf1d90> <<< 13118 1727204065.60185: stdout chunk (state=3): >>># extension module '_blake2' loaded from '/usr/lib64/python3.9/lib-dynload/_blake2.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204065.60188: stdout chunk (state=3): >>># extension module '_blake2' executed from '/usr/lib64/python3.9/lib-dynload/_blake2.cpython-39-x86_64-linux-gnu.so' import '_blake2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cfb2e0> import 'hashlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0cf19d0> <<< 13118 1727204065.60206: stdout chunk (state=3): >>>import 'random' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0ce5b20> <<< 13118 1727204065.60228: stdout chunk (state=3): >>>import 'tempfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10416a0> <<< 13118 1727204065.60254: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/zipfile.cpython-39.pyc matches /usr/lib64/python3.9/zipfile.py <<< 13118 1727204065.60310: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/zipfile.cpython-39.pyc' <<< 13118 1727204065.60347: stdout chunk (state=3): >>>import 'zipfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0cf1b80> <<< 13118 1727204065.60447: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/encodings/cp437.pyc' <<< 13118 1727204065.60460: stdout chunk (state=3): >>>import 'encodings.cp437' # <_frozen_importlib_external.SourcelessFileLoader object at 0x7ff7c0c14760> <<< 13118 1727204065.60594: stdout chunk (state=3): >>># zipimport: found 30 names in '/tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip' <<< 13118 1727204065.60597: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.60678: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.60710: stdout chunk (state=3): >>>import ansible # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/__init__.py <<< 13118 1727204065.60714: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.60742: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.60745: stdout chunk (state=3): >>>import ansible.module_utils # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/__init__.py <<< 13118 1727204065.60758: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.62004: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.62961: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/__future__.cpython-39.pyc matches /usr/lib64/python3.9/__future__.py # code object from '/usr/lib64/python3.9/__pycache__/__future__.cpython-39.pyc' import '__future__' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b3b8b0> <<< 13118 1727204065.62980: stdout chunk (state=3): >>># /usr/lib64/python3.9/json/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/json/__init__.py # code object from '/usr/lib64/python3.9/json/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204065.63007: stdout chunk (state=3): >>># /usr/lib64/python3.9/json/__pycache__/decoder.cpython-39.pyc matches /usr/lib64/python3.9/json/decoder.py # code object from '/usr/lib64/python3.9/json/__pycache__/decoder.cpython-39.pyc' <<< 13118 1727204065.63038: stdout chunk (state=3): >>># /usr/lib64/python3.9/json/__pycache__/scanner.cpython-39.pyc matches /usr/lib64/python3.9/json/scanner.py # code object from '/usr/lib64/python3.9/json/__pycache__/scanner.cpython-39.pyc' <<< 13118 1727204065.63068: stdout chunk (state=3): >>># extension module '_json' loaded from '/usr/lib64/python3.9/lib-dynload/_json.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204065.63071: stdout chunk (state=3): >>># extension module '_json' executed from '/usr/lib64/python3.9/lib-dynload/_json.cpython-39-x86_64-linux-gnu.so' import '_json' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0b3b160> <<< 13118 1727204065.63109: stdout chunk (state=3): >>>import 'json.scanner' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b3b280> <<< 13118 1727204065.63147: stdout chunk (state=3): >>>import 'json.decoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b3b5e0> <<< 13118 1727204065.63162: stdout chunk (state=3): >>># /usr/lib64/python3.9/json/__pycache__/encoder.cpython-39.pyc matches /usr/lib64/python3.9/json/encoder.py # code object from '/usr/lib64/python3.9/json/__pycache__/encoder.cpython-39.pyc' <<< 13118 1727204065.63210: stdout chunk (state=3): >>>import 'json.encoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b3b4f0> import 'json' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b3be20> <<< 13118 1727204065.63222: stdout chunk (state=3): >>>import 'atexit' # <<< 13118 1727204065.63261: stdout chunk (state=3): >>># extension module 'fcntl' loaded from '/usr/lib64/python3.9/lib-dynload/fcntl.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204065.63266: stdout chunk (state=3): >>># extension module 'fcntl' executed from '/usr/lib64/python3.9/lib-dynload/fcntl.cpython-39-x86_64-linux-gnu.so' import 'fcntl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0b3b580> <<< 13118 1727204065.63276: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/locale.cpython-39.pyc matches /usr/lib64/python3.9/locale.py <<< 13118 1727204065.63307: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/locale.cpython-39.pyc' <<< 13118 1727204065.63359: stdout chunk (state=3): >>>import 'locale' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b3b100> <<< 13118 1727204065.63364: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/platform.cpython-39.pyc matches /usr/lib64/python3.9/platform.py <<< 13118 1727204065.63380: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/platform.cpython-39.pyc' <<< 13118 1727204065.63394: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/subprocess.cpython-39.pyc matches /usr/lib64/python3.9/subprocess.py <<< 13118 1727204065.63415: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/subprocess.cpython-39.pyc' <<< 13118 1727204065.63444: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/signal.cpython-39.pyc matches /usr/lib64/python3.9/signal.py # code object from '/usr/lib64/python3.9/__pycache__/signal.cpython-39.pyc' <<< 13118 1727204065.63515: stdout chunk (state=3): >>>import 'signal' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c056dfd0> <<< 13118 1727204065.63553: stdout chunk (state=3): >>># extension module '_posixsubprocess' loaded from '/usr/lib64/python3.9/lib-dynload/_posixsubprocess.cpython-39-x86_64-linux-gnu.so' # extension module '_posixsubprocess' executed from '/usr/lib64/python3.9/lib-dynload/_posixsubprocess.cpython-39-x86_64-linux-gnu.so' import '_posixsubprocess' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c058bc40> <<< 13118 1727204065.63589: stdout chunk (state=3): >>># extension module 'select' loaded from '/usr/lib64/python3.9/lib-dynload/select.cpython-39-x86_64-linux-gnu.so' # extension module 'select' executed from '/usr/lib64/python3.9/lib-dynload/select.cpython-39-x86_64-linux-gnu.so' import 'select' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c058bf40> <<< 13118 1727204065.63602: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/selectors.cpython-39.pyc matches /usr/lib64/python3.9/selectors.py <<< 13118 1727204065.63634: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/selectors.cpython-39.pyc' <<< 13118 1727204065.63670: stdout chunk (state=3): >>>import 'selectors' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c058b2e0> <<< 13118 1727204065.63688: stdout chunk (state=3): >>>import 'subprocess' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0ba3d90> <<< 13118 1727204065.63852: stdout chunk (state=3): >>>import 'platform' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0ba33a0> <<< 13118 1727204065.63878: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/shlex.cpython-39.pyc matches /usr/lib64/python3.9/shlex.py <<< 13118 1727204065.63881: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/shlex.cpython-39.pyc' <<< 13118 1727204065.63914: stdout chunk (state=3): >>>import 'shlex' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0ba3f40> <<< 13118 1727204065.63917: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/traceback.cpython-39.pyc matches /usr/lib64/python3.9/traceback.py <<< 13118 1727204065.63941: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/traceback.cpython-39.pyc' <<< 13118 1727204065.63956: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/linecache.cpython-39.pyc matches /usr/lib64/python3.9/linecache.py # code object from '/usr/lib64/python3.9/__pycache__/linecache.cpython-39.pyc' <<< 13118 1727204065.63980: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/tokenize.cpython-39.pyc matches /usr/lib64/python3.9/tokenize.py <<< 13118 1727204065.63992: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/tokenize.cpython-39.pyc' <<< 13118 1727204065.64022: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/token.cpython-39.pyc matches /usr/lib64/python3.9/token.py <<< 13118 1727204065.64025: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/token.cpython-39.pyc' <<< 13118 1727204065.64028: stdout chunk (state=3): >>>import 'token' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0c14a90> <<< 13118 1727204065.64113: stdout chunk (state=3): >>>import 'tokenize' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05e9dc0> <<< 13118 1727204065.64117: stdout chunk (state=3): >>>import 'linecache' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05e9490> import 'traceback' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b45580> <<< 13118 1727204065.64175: stdout chunk (state=3): >>># extension module 'syslog' loaded from '/usr/lib64/python3.9/lib-dynload/syslog.cpython-39-x86_64-linux-gnu.so' # extension module 'syslog' executed from '/usr/lib64/python3.9/lib-dynload/syslog.cpython-39-x86_64-linux-gnu.so' import 'syslog' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c05e95b0> <<< 13118 1727204065.64190: stdout chunk (state=3): >>># /usr/lib64/python3.9/site-packages/systemd/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/__init__.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/__init__.cpython-39.pyc' import 'systemd' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05e95e0> <<< 13118 1727204065.64207: stdout chunk (state=3): >>># /usr/lib64/python3.9/site-packages/systemd/__pycache__/journal.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/journal.py <<< 13118 1727204065.64218: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/journal.cpython-39.pyc' <<< 13118 1727204065.64243: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/datetime.cpython-39.pyc matches /usr/lib64/python3.9/datetime.py <<< 13118 1727204065.64277: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/datetime.cpython-39.pyc' <<< 13118 1727204065.64339: stdout chunk (state=3): >>># extension module '_datetime' loaded from '/usr/lib64/python3.9/lib-dynload/_datetime.cpython-39-x86_64-linux-gnu.so' # extension module '_datetime' executed from '/usr/lib64/python3.9/lib-dynload/_datetime.cpython-39-x86_64-linux-gnu.so' import '_datetime' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c055ef70> import 'datetime' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b832e0> <<< 13118 1727204065.64386: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/uuid.cpython-39.pyc matches /usr/lib64/python3.9/uuid.py # code object from '/usr/lib64/python3.9/__pycache__/uuid.cpython-39.pyc' <<< 13118 1727204065.64442: stdout chunk (state=3): >>># extension module '_uuid' loaded from '/usr/lib64/python3.9/lib-dynload/_uuid.cpython-39-x86_64-linux-gnu.so' # extension module '_uuid' executed from '/usr/lib64/python3.9/lib-dynload/_uuid.cpython-39-x86_64-linux-gnu.so' import '_uuid' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c055b7f0> import 'uuid' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b83460> <<< 13118 1727204065.64461: stdout chunk (state=3): >>># /usr/lib64/python3.9/logging/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/logging/__init__.py <<< 13118 1727204065.64514: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/logging/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204065.64517: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/string.cpython-39.pyc matches /usr/lib64/python3.9/string.py <<< 13118 1727204065.64531: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/string.cpython-39.pyc' import '_string' # <<< 13118 1727204065.64595: stdout chunk (state=3): >>>import 'string' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b9bf40> <<< 13118 1727204065.64724: stdout chunk (state=3): >>>import 'logging' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c055b790> <<< 13118 1727204065.64853: stdout chunk (state=3): >>># extension module 'systemd._journal' loaded from '/usr/lib64/python3.9/site-packages/systemd/_journal.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._journal' executed from '/usr/lib64/python3.9/site-packages/systemd/_journal.cpython-39-x86_64-linux-gnu.so' import 'systemd._journal' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c055b5e0> <<< 13118 1727204065.64856: stdout chunk (state=3): >>># extension module 'systemd._reader' loaded from '/usr/lib64/python3.9/site-packages/systemd/_reader.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._reader' executed from '/usr/lib64/python3.9/site-packages/systemd/_reader.cpython-39-x86_64-linux-gnu.so' import 'systemd._reader' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c055a550> <<< 13118 1727204065.64896: stdout chunk (state=3): >>># extension module 'systemd.id128' loaded from '/usr/lib64/python3.9/site-packages/systemd/id128.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204065.64899: stdout chunk (state=3): >>># extension module 'systemd.id128' executed from '/usr/lib64/python3.9/site-packages/systemd/id128.cpython-39-x86_64-linux-gnu.so' import 'systemd.id128' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c055a490> import 'systemd.journal' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b7c9a0> <<< 13118 1727204065.64927: stdout chunk (state=3): >>># /usr/lib64/python3.9/site-packages/systemd/__pycache__/daemon.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/daemon.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/daemon.cpython-39.pyc' <<< 13118 1727204065.64952: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/socket.cpython-39.pyc matches /usr/lib64/python3.9/socket.py <<< 13118 1727204065.64973: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/socket.cpython-39.pyc' <<< 13118 1727204065.65012: stdout chunk (state=3): >>># extension module '_socket' loaded from '/usr/lib64/python3.9/lib-dynload/_socket.cpython-39-x86_64-linux-gnu.so' # extension module '_socket' executed from '/usr/lib64/python3.9/lib-dynload/_socket.cpython-39-x86_64-linux-gnu.so' import '_socket' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c05df6a0> <<< 13118 1727204065.65286: stdout chunk (state=3): >>># extension module 'array' loaded from '/usr/lib64/python3.9/lib-dynload/array.cpython-39-x86_64-linux-gnu.so' # extension module 'array' executed from '/usr/lib64/python3.9/lib-dynload/array.cpython-39-x86_64-linux-gnu.so' import 'array' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c05debb0> import 'socket' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05ef0d0> # extension module 'systemd._daemon' loaded from '/usr/lib64/python3.9/site-packages/systemd/_daemon.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._daemon' executed from '/usr/lib64/python3.9/site-packages/systemd/_daemon.cpython-39-x86_64-linux-gnu.so' import 'systemd._daemon' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c05df100> import 'systemd.daemon' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b47c40> # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.compat # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/compat/__init__.py <<< 13118 1727204065.65298: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.65404: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.65439: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.65451: stdout chunk (state=3): >>># zipimport: zlib available import ansible.module_utils.common # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/__init__.py <<< 13118 1727204065.65482: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13118 1727204065.65509: stdout chunk (state=3): >>>import ansible.module_utils.common.text # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/text/__init__.py <<< 13118 1727204065.65513: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.65610: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.65702: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.66158: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.66628: stdout chunk (state=3): >>>import ansible.module_utils.six # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/six/__init__.py <<< 13118 1727204065.66632: stdout chunk (state=3): >>>import 'ansible.module_utils.six.moves' # import 'ansible.module_utils.six.moves.collections_abc' # <<< 13118 1727204065.66634: stdout chunk (state=3): >>>import ansible.module_utils.common.text.converters # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/text/converters.py <<< 13118 1727204065.66657: stdout chunk (state=3): >>># /usr/lib64/python3.9/ctypes/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/ctypes/__init__.py <<< 13118 1727204065.66672: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/ctypes/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204065.66735: stdout chunk (state=3): >>># extension module '_ctypes' loaded from '/usr/lib64/python3.9/lib-dynload/_ctypes.cpython-39-x86_64-linux-gnu.so' <<< 13118 1727204065.66738: stdout chunk (state=3): >>># extension module '_ctypes' executed from '/usr/lib64/python3.9/lib-dynload/_ctypes.cpython-39-x86_64-linux-gnu.so' import '_ctypes' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0526940> <<< 13118 1727204065.66804: stdout chunk (state=3): >>># /usr/lib64/python3.9/ctypes/__pycache__/_endian.cpython-39.pyc matches /usr/lib64/python3.9/ctypes/_endian.py # code object from '/usr/lib64/python3.9/ctypes/__pycache__/_endian.cpython-39.pyc' <<< 13118 1727204065.66810: stdout chunk (state=3): >>>import 'ctypes._endian' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05dcd30> <<< 13118 1727204065.66822: stdout chunk (state=3): >>>import 'ctypes' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05d37c0> <<< 13118 1727204065.66881: stdout chunk (state=3): >>>import ansible.module_utils.compat.selinux # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/compat/selinux.py <<< 13118 1727204065.66885: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.66900: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.66917: stdout chunk (state=3): >>>import ansible.module_utils._text # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/_text.py <<< 13118 1727204065.66920: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.67061: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.67181: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/copy.cpython-39.pyc matches /usr/lib64/python3.9/copy.py <<< 13118 1727204065.67184: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/copy.cpython-39.pyc' <<< 13118 1727204065.67200: stdout chunk (state=3): >>>import 'copy' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05de4c0> <<< 13118 1727204065.67218: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.67599: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.67974: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.68027: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.68096: stdout chunk (state=3): >>>import ansible.module_utils.common.collections # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/collections.py <<< 13118 1727204065.68099: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.68138: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.68168: stdout chunk (state=3): >>>import ansible.module_utils.common.warnings # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/warnings.py <<< 13118 1727204065.68180: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.68242: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.68311: stdout chunk (state=3): >>>import ansible.module_utils.errors # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/errors.py <<< 13118 1727204065.68316: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.68330: stdout chunk (state=3): >>># zipimport: zlib available import ansible.module_utils.parsing # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/parsing/__init__.py <<< 13118 1727204065.68358: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.68386: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.68426: stdout chunk (state=3): >>>import ansible.module_utils.parsing.convert_bool # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/parsing/convert_bool.py <<< 13118 1727204065.68432: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.68618: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.68807: stdout chunk (state=3): >>># /usr/lib64/python3.9/__pycache__/ast.cpython-39.pyc matches /usr/lib64/python3.9/ast.py <<< 13118 1727204065.68845: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/__pycache__/ast.cpython-39.pyc' <<< 13118 1727204065.68848: stdout chunk (state=3): >>>import '_ast' # <<< 13118 1727204065.68920: stdout chunk (state=3): >>>import 'ast' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c00be940> <<< 13118 1727204065.68924: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.68985: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.69061: stdout chunk (state=3): >>>import ansible.module_utils.common.text.formatters # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/text/formatters.py <<< 13118 1727204065.69067: stdout chunk (state=3): >>>import ansible.module_utils.common.validation # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/validation.py import ansible.module_utils.common.parameters # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/parameters.py <<< 13118 1727204065.69071: stdout chunk (state=3): >>>import ansible.module_utils.common.arg_spec # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/arg_spec.py <<< 13118 1727204065.69086: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.69129: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.69162: stdout chunk (state=3): >>>import ansible.module_utils.common.locale # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/locale.py <<< 13118 1727204065.69169: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.69209: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.69246: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.69367: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.69402: stdout chunk (state=3): >>># /usr/lib64/python3.9/site-packages/selinux/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/selinux/__init__.py <<< 13118 1727204065.69430: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.9/site-packages/selinux/__pycache__/__init__.cpython-39.pyc' <<< 13118 1727204065.69515: stdout chunk (state=3): >>># extension module 'selinux._selinux' loaded from '/usr/lib64/python3.9/site-packages/selinux/_selinux.cpython-39-x86_64-linux-gnu.so' # extension module 'selinux._selinux' executed from '/usr/lib64/python3.9/site-packages/selinux/_selinux.cpython-39-x86_64-linux-gnu.so' import 'selinux._selinux' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0b8eb50> <<< 13118 1727204065.69535: stdout chunk (state=3): >>>import 'selinux' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c00bd070> <<< 13118 1727204065.69569: stdout chunk (state=3): >>>import ansible.module_utils.common.file # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/file.py import ansible.module_utils.common.process # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/process.py <<< 13118 1727204065.69592: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.70291: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib/python3.9/site-packages/__pycache__/distro.cpython-39.pyc matches /usr/lib/python3.9/site-packages/distro.py # code object from '/usr/lib/python3.9/site-packages/__pycache__/distro.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/argparse.cpython-39.pyc matches /usr/lib64/python3.9/argparse.py # code object from '/usr/lib64/python3.9/__pycache__/argparse.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/gettext.cpython-39.pyc matches /usr/lib64/python3.9/gettext.py # code object from '/usr/lib64/python3.9/__pycache__/gettext.cpython-39.pyc' import 'gettext' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c010e6d0> import 'argparse' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c051dc10> import 'distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c051c5b0> # destroy ansible.module_utils.distro import ansible.module_utils.distro # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/distro/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common._utils # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/_utils.py import ansible.module_utils.common.sys_info # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/sys_info.py <<< 13118 1727204065.70405: stdout chunk (state=3): >>>import ansible.module_utils.basic # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/basic.py <<< 13118 1727204065.70426: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13118 1727204065.70432: stdout chunk (state=3): >>>import ansible.modules # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/modules/__init__.py # zipimport: zlib available <<< 13118 1727204065.70608: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.70873: stdout chunk (state=3): >>># zipimport: zlib available <<< 13118 1727204065.71166: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/run/ostree-booted", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} <<< 13118 1727204065.71199: stdout chunk (state=3): >>># destroy __main__ <<< 13118 1727204065.71616: stdout chunk (state=3): >>># clear builtins._ # clear sys.path <<< 13118 1727204065.71624: stdout chunk (state=3): >>># clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_type # clear sys.last_value<<< 13118 1727204065.71712: stdout chunk (state=3): >>> # clear sys.last_traceback # clear sys.path_hooks # clear sys.path_importer_cache <<< 13118 1727204065.71849: stdout chunk (state=3): >>># clear sys.meta_path # clear sys.__interactivehook__ <<< 13118 1727204065.72282: stdout chunk (state=3): >>># restore sys.stdin <<< 13118 1727204065.72354: stdout chunk (state=3): >>># restore sys.stdout # restore sys.stderr<<< 13118 1727204065.72361: stdout chunk (state=3): >>> <<< 13118 1727204065.72367: stdout chunk (state=3): >>># cleanup[2] removing sys <<< 13118 1727204065.72370: stdout chunk (state=3): >>># cleanup[2] removing builtins <<< 13118 1727204065.72373: stdout chunk (state=3): >>># cleanup[2] removing _frozen_importlib <<< 13118 1727204065.72375: stdout chunk (state=3): >>># cleanup[2] removing _imp <<< 13118 1727204065.72379: stdout chunk (state=3): >>># cleanup[2] removing _thread # cleanup[2] removing _warnings<<< 13118 1727204065.72383: stdout chunk (state=3): >>> # cleanup[2] removing _weakref<<< 13118 1727204065.72385: stdout chunk (state=3): >>> <<< 13118 1727204065.72389: stdout chunk (state=3): >>># cleanup[2] removing _io <<< 13118 1727204065.72469: stdout chunk (state=3): >>># cleanup[2] removing marshal<<< 13118 1727204065.72473: stdout chunk (state=3): >>> # cleanup[2] removing posix <<< 13118 1727204065.72475: stdout chunk (state=3): >>># cleanup[2] removing _frozen_importlib_external<<< 13118 1727204065.72477: stdout chunk (state=3): >>> # cleanup[2] removing time <<< 13118 1727204065.72483: stdout chunk (state=3): >>># cleanup[2] removing zipimport<<< 13118 1727204065.72485: stdout chunk (state=3): >>> # cleanup[2] removing _codecs<<< 13118 1727204065.72487: stdout chunk (state=3): >>> <<< 13118 1727204065.72489: stdout chunk (state=3): >>># cleanup[2] removing codecs # cleanup[2] removing encodings.aliases<<< 13118 1727204065.72491: stdout chunk (state=3): >>> <<< 13118 1727204065.72492: stdout chunk (state=3): >>># cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8<<< 13118 1727204065.72494: stdout chunk (state=3): >>> # cleanup[2] removing _signal<<< 13118 1727204065.72496: stdout chunk (state=3): >>> # cleanup[2] removing encodings.latin_1 <<< 13118 1727204065.72497: stdout chunk (state=3): >>># cleanup[2] removing _abc<<< 13118 1727204065.72499: stdout chunk (state=3): >>> # cleanup[2] removing abc <<< 13118 1727204065.72501: stdout chunk (state=3): >>># cleanup[2] removing io # cleanup[2] removing __main__ <<< 13118 1727204065.72503: stdout chunk (state=3): >>># cleanup[2] removing _stat <<< 13118 1727204065.72505: stdout chunk (state=3): >>># cleanup[2] removing stat <<< 13118 1727204065.72507: stdout chunk (state=3): >>># cleanup[2] removing _collections_abc<<< 13118 1727204065.72508: stdout chunk (state=3): >>> # cleanup[2] removing genericpath <<< 13118 1727204065.72510: stdout chunk (state=3): >>># cleanup[2] removing posixpath <<< 13118 1727204065.72512: stdout chunk (state=3): >>># cleanup[2] removing os.path <<< 13118 1727204065.72514: stdout chunk (state=3): >>># cleanup[2] removing os<<< 13118 1727204065.72516: stdout chunk (state=3): >>> # cleanup[2] removing _sitebuiltins<<< 13118 1727204065.72518: stdout chunk (state=3): >>> <<< 13118 1727204065.72520: stdout chunk (state=3): >>># cleanup[2] removing _locale<<< 13118 1727204065.72522: stdout chunk (state=3): >>> # cleanup[2] removing _bootlocale<<< 13118 1727204065.72524: stdout chunk (state=3): >>> <<< 13118 1727204065.72526: stdout chunk (state=3): >>># destroy _bootlocale <<< 13118 1727204065.72528: stdout chunk (state=3): >>># cleanup[2] removing site <<< 13118 1727204065.72529: stdout chunk (state=3): >>># destroy site<<< 13118 1727204065.72531: stdout chunk (state=3): >>> <<< 13118 1727204065.72533: stdout chunk (state=3): >>># cleanup[2] removing types <<< 13118 1727204065.72535: stdout chunk (state=3): >>># cleanup[2] removing enum # cleanup[2] removing _sre<<< 13118 1727204065.72537: stdout chunk (state=3): >>> # cleanup[2] removing sre_constants<<< 13118 1727204065.72539: stdout chunk (state=3): >>> <<< 13118 1727204065.72541: stdout chunk (state=3): >>># destroy sre_constants <<< 13118 1727204065.72543: stdout chunk (state=3): >>># cleanup[2] removing sre_parse <<< 13118 1727204065.72545: stdout chunk (state=3): >>># cleanup[2] removing sre_compile # cleanup[2] removing _heapq<<< 13118 1727204065.72546: stdout chunk (state=3): >>> # cleanup[2] removing heapq<<< 13118 1727204065.72548: stdout chunk (state=3): >>> # cleanup[2] removing itertools <<< 13118 1727204065.72550: stdout chunk (state=3): >>># cleanup[2] removing keyword<<< 13118 1727204065.72552: stdout chunk (state=3): >>> # destroy keyword <<< 13118 1727204065.72553: stdout chunk (state=3): >>># cleanup[2] removing _operator<<< 13118 1727204065.72556: stdout chunk (state=3): >>> # cleanup[2] removing operator<<< 13118 1727204065.72558: stdout chunk (state=3): >>> # cleanup[2] removing reprlib<<< 13118 1727204065.72561: stdout chunk (state=3): >>> # destroy reprlib <<< 13118 1727204065.72563: stdout chunk (state=3): >>># cleanup[2] removing _collections <<< 13118 1727204065.72567: stdout chunk (state=3): >>># cleanup[2] removing collections<<< 13118 1727204065.72569: stdout chunk (state=3): >>> <<< 13118 1727204065.72572: stdout chunk (state=3): >>># cleanup[2] removing _functools <<< 13118 1727204065.72574: stdout chunk (state=3): >>># cleanup[2] removing functools <<< 13118 1727204065.72576: stdout chunk (state=3): >>># cleanup[2] removing copyreg<<< 13118 1727204065.72583: stdout chunk (state=3): >>> # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # destroy base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing collections.abc # cleanup[2] removing contextlib # cleanup[2] removing typing # destroy typing # cleanup[2] removing importlib.abc # cleanup[2] removing importlib.util # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing threading # cleanup[2] removing _bz2 # destroy _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing pwd # cleanup[2] removing grp # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing tempfile # cleanup[2] removing zipfile # destroy zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json <<< 13118 1727204065.72586: stdout chunk (state=3): >>># cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing fcntl # cleanup[2] removing locale # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing subprocess # cleanup[2] removing platform # cleanup[2] removing shlex # cleanup[2] removing token # destroy token # cleanup[2] removing tokenize <<< 13118 1727204065.72589: stdout chunk (state=3): >>># cleanup[2] removing linecache # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # destroy socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves<<< 13118 1727204065.72591: stdout chunk (state=3): >>> # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian <<< 13118 1727204065.72594: stdout chunk (state=3): >>># cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool <<< 13118 1727204065.72600: stdout chunk (state=3): >>># cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters <<< 13118 1727204065.72602: stdout chunk (state=3): >>># destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters<<< 13118 1727204065.72604: stdout chunk (state=3): >>> # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules <<< 13118 1727204065.72785: stdout chunk (state=3): >>># destroy _sitebuiltins # destroy importlib.util <<< 13118 1727204065.72843: stdout chunk (state=3): >>># destroy importlib.abc # destroy importlib.machinery <<< 13118 1727204065.73097: stdout chunk (state=3): >>># destroy zipimport # destroy _compression # destroy binascii # destroy importlib # destroy struct # destroy bz2 # destroy lzma # destroy __main__ # destroy locale # destroy tempfile # destroy systemd.journal # destroy systemd.daemon # destroy ansible.module_utils.compat.selinux # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy encodings # destroy syslog # destroy uuid # destroy array # destroy datetime <<< 13118 1727204065.73247: stdout chunk (state=3): >>># destroy selinux # destroy distro # destroy json # destroy shlex # destroy logging # destroy argparse # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # cleanup[3] wiping tokenize # cleanup[3] wiping platform # destroy subprocess # cleanup[3] wiping selectors # cleanup[3] wiping select # cleanup[3] wiping _posixsubprocess # cleanup[3] wiping signal # cleanup[3] wiping fcntl # cleanup[3] wiping atexit # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping _blake2 # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping shutil # destroy fnmatch # cleanup[3] wiping grp # cleanup[3] wiping pwd # cleanup[3] wiping _lzma # cleanup[3] wiping threading # cleanup[3] wiping zlib # cleanup[3] wiping errno # cleanup[3] wiping weakref # cleanup[3] wiping contextlib # cleanup[3] wiping collections.abc # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy enum # destroy sre_compile # destroy copyreg # cleanup[3] wiping functools # cleanup[3] wiping _functools # destroy _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy heapq # destroy collections.abc # cleanup[3] wiping _collections # destroy _collections # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping itertools # cleanup[3] wiping _heapq # cleanup[3] wiping sre_parse # cleanup[3] wiping _sre # cleanup[3] wiping types # cleanup[3] wiping _locale # destroy _locale # cleanup[3] wiping os # cleanup[3] wiping os.path # destroy genericpath # cleanup[3] wiping posixpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.latin_1 # cleanup[3] wiping _signal # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs <<< 13118 1727204065.73278: stdout chunk (state=3): >>># cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy systemd._daemon # destroy _socket # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy fcntl # destroy _blake2 # destroy _lzma # destroy zlib # destroy _signal <<< 13118 1727204065.73426: stdout chunk (state=3): >>># destroy platform # destroy _uuid # destroy _sre # destroy sre_parse<<< 13118 1727204065.73448: stdout chunk (state=3): >>> # destroy tokenize <<< 13118 1727204065.73461: stdout chunk (state=3): >>># destroy _heapq # destroy posixpath # destroy stat <<< 13118 1727204065.73492: stdout chunk (state=3): >>># destroy ansible.module_utils.six.moves.urllib # destroy errno # destroy signal # destroy contextlib # destroy pwd<<< 13118 1727204065.73496: stdout chunk (state=3): >>> # destroy grp # destroy _posixsubprocess # destroy selectors <<< 13118 1727204065.73499: stdout chunk (state=3): >>># destroy select # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser<<< 13118 1727204065.73502: stdout chunk (state=3): >>> # destroy functools # destroy itertools # destroy operator # destroy ansible.module_utils.six.moves # destroy _operator # destroy _frozen_importlib_external # destroy _imp # destroy io<<< 13118 1727204065.73505: stdout chunk (state=3): >>> # destroy marshal <<< 13118 1727204065.73773: stdout chunk (state=3): >>># destroy _frozen_importlib<<< 13118 1727204065.73777: stdout chunk (state=3): >>> # clear sys.audit hooks <<< 13118 1727204065.73910: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204065.73913: stdout chunk (state=3): >>><<< 13118 1727204065.73915: stderr chunk (state=3): >>><<< 13118 1727204065.74079: _low_level_execute_command() done: rc=0, stdout=import _frozen_importlib # frozen import _imp # builtin import '_thread' # import '_warnings' # import '_weakref' # import '_io' # import 'marshal' # import 'posix' # import '_frozen_importlib_external' # # installing zipimport hook import 'time' # import 'zipimport' # # installed zipimport hook # /usr/lib64/python3.9/encodings/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/encodings/__init__.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/codecs.cpython-39.pyc matches /usr/lib64/python3.9/codecs.py # code object from '/usr/lib64/python3.9/__pycache__/codecs.cpython-39.pyc' import '_codecs' # import 'codecs' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c11b3dc0> # /usr/lib64/python3.9/encodings/__pycache__/aliases.cpython-39.pyc matches /usr/lib64/python3.9/encodings/aliases.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/aliases.cpython-39.pyc' import 'encodings.aliases' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c11583a0> import 'encodings' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c11b3b20> # /usr/lib64/python3.9/encodings/__pycache__/utf_8.cpython-39.pyc matches /usr/lib64/python3.9/encodings/utf_8.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/utf_8.cpython-39.pyc' import 'encodings.utf_8' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c11b3ac0> import '_signal' # # /usr/lib64/python3.9/encodings/__pycache__/latin_1.cpython-39.pyc matches /usr/lib64/python3.9/encodings/latin_1.py # code object from '/usr/lib64/python3.9/encodings/__pycache__/latin_1.cpython-39.pyc' import 'encodings.latin_1' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1158490> # /usr/lib64/python3.9/__pycache__/io.cpython-39.pyc matches /usr/lib64/python3.9/io.py # code object from '/usr/lib64/python3.9/__pycache__/io.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/abc.py # code object from '/usr/lib64/python3.9/__pycache__/abc.cpython-39.pyc' import '_abc' # import 'abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1158940> import 'io' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1158670> # /usr/lib64/python3.9/__pycache__/site.cpython-39.pyc matches /usr/lib64/python3.9/site.py # code object from '/usr/lib64/python3.9/__pycache__/site.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/os.cpython-39.pyc matches /usr/lib64/python3.9/os.py # code object from '/usr/lib64/python3.9/__pycache__/os.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/stat.cpython-39.pyc matches /usr/lib64/python3.9/stat.py # code object from '/usr/lib64/python3.9/__pycache__/stat.cpython-39.pyc' import '_stat' # import 'stat' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c110f190> # /usr/lib64/python3.9/__pycache__/_collections_abc.cpython-39.pyc matches /usr/lib64/python3.9/_collections_abc.py # code object from '/usr/lib64/python3.9/__pycache__/_collections_abc.cpython-39.pyc' import '_collections_abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c110f220> # /usr/lib64/python3.9/__pycache__/posixpath.cpython-39.pyc matches /usr/lib64/python3.9/posixpath.py # code object from '/usr/lib64/python3.9/__pycache__/posixpath.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/genericpath.cpython-39.pyc matches /usr/lib64/python3.9/genericpath.py # code object from '/usr/lib64/python3.9/__pycache__/genericpath.cpython-39.pyc' import 'genericpath' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1132850> import 'posixpath' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c110f940> import 'os' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1170880> # /usr/lib64/python3.9/__pycache__/_sitebuiltins.cpython-39.pyc matches /usr/lib64/python3.9/_sitebuiltins.py # code object from '/usr/lib64/python3.9/__pycache__/_sitebuiltins.cpython-39.pyc' import '_sitebuiltins' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1108d90> # /usr/lib64/python3.9/__pycache__/_bootlocale.cpython-39.pyc matches /usr/lib64/python3.9/_bootlocale.py # code object from '/usr/lib64/python3.9/__pycache__/_bootlocale.cpython-39.pyc' import '_locale' # import '_bootlocale' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1132d90> import 'site' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1158970> Python 3.9.19 (main, Aug 23 2024, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-2)] on linux Type "help", "copyright", "credits" or "license" for more information. # /usr/lib64/python3.9/__pycache__/base64.cpython-39.pyc matches /usr/lib64/python3.9/base64.py # code object from '/usr/lib64/python3.9/__pycache__/base64.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/re.cpython-39.pyc matches /usr/lib64/python3.9/re.py # code object from '/usr/lib64/python3.9/__pycache__/re.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/enum.cpython-39.pyc matches /usr/lib64/python3.9/enum.py # code object from '/usr/lib64/python3.9/__pycache__/enum.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/types.cpython-39.pyc matches /usr/lib64/python3.9/types.py # code object from '/usr/lib64/python3.9/__pycache__/types.cpython-39.pyc' import 'types' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10adf10> import 'enum' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10b40a0> # /usr/lib64/python3.9/__pycache__/sre_compile.cpython-39.pyc matches /usr/lib64/python3.9/sre_compile.py # code object from '/usr/lib64/python3.9/__pycache__/sre_compile.cpython-39.pyc' import '_sre' # # /usr/lib64/python3.9/__pycache__/sre_parse.cpython-39.pyc matches /usr/lib64/python3.9/sre_parse.py # code object from '/usr/lib64/python3.9/__pycache__/sre_parse.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/sre_constants.cpython-39.pyc matches /usr/lib64/python3.9/sre_constants.py # code object from '/usr/lib64/python3.9/__pycache__/sre_constants.cpython-39.pyc' import 'sre_constants' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10a75b0> import 'sre_parse' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10ae6a0> import 'sre_compile' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10ad3d0> # /usr/lib64/python3.9/__pycache__/functools.cpython-39.pyc matches /usr/lib64/python3.9/functools.py # code object from '/usr/lib64/python3.9/__pycache__/functools.cpython-39.pyc' # /usr/lib64/python3.9/collections/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/collections/__init__.py # code object from '/usr/lib64/python3.9/collections/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/heapq.cpython-39.pyc matches /usr/lib64/python3.9/heapq.py # code object from '/usr/lib64/python3.9/__pycache__/heapq.cpython-39.pyc' # extension module '_heapq' loaded from '/usr/lib64/python3.9/lib-dynload/_heapq.cpython-39-x86_64-linux-gnu.so' # extension module '_heapq' executed from '/usr/lib64/python3.9/lib-dynload/_heapq.cpython-39-x86_64-linux-gnu.so' import '_heapq' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c1031eb0> import 'heapq' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10319a0> import 'itertools' # # /usr/lib64/python3.9/__pycache__/keyword.cpython-39.pyc matches /usr/lib64/python3.9/keyword.py # code object from '/usr/lib64/python3.9/__pycache__/keyword.cpython-39.pyc' import 'keyword' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1031fa0> # /usr/lib64/python3.9/__pycache__/operator.cpython-39.pyc matches /usr/lib64/python3.9/operator.py # code object from '/usr/lib64/python3.9/__pycache__/operator.cpython-39.pyc' import '_operator' # import 'operator' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1031df0> # /usr/lib64/python3.9/__pycache__/reprlib.cpython-39.pyc matches /usr/lib64/python3.9/reprlib.py # code object from '/usr/lib64/python3.9/__pycache__/reprlib.cpython-39.pyc' import 'reprlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1041160> import '_collections' # import 'collections' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1089e20> import '_functools' # import 'functools' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1081700> # /usr/lib64/python3.9/__pycache__/copyreg.cpython-39.pyc matches /usr/lib64/python3.9/copyreg.py # code object from '/usr/lib64/python3.9/__pycache__/copyreg.cpython-39.pyc' import 'copyreg' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1095760> import 're' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10b5eb0> # /usr/lib64/python3.9/__pycache__/struct.cpython-39.pyc matches /usr/lib64/python3.9/struct.py # code object from '/usr/lib64/python3.9/__pycache__/struct.cpython-39.pyc' # extension module '_struct' loaded from '/usr/lib64/python3.9/lib-dynload/_struct.cpython-39-x86_64-linux-gnu.so' # extension module '_struct' executed from '/usr/lib64/python3.9/lib-dynload/_struct.cpython-39-x86_64-linux-gnu.so' import '_struct' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c1041d60> import 'struct' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1089340> # extension module 'binascii' loaded from '/usr/lib64/python3.9/lib-dynload/binascii.cpython-39-x86_64-linux-gnu.so' # extension module 'binascii' executed from '/usr/lib64/python3.9/lib-dynload/binascii.cpython-39-x86_64-linux-gnu.so' import 'binascii' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c1095370> import 'base64' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10bba60> # /usr/lib64/python3.9/__pycache__/runpy.cpython-39.pyc matches /usr/lib64/python3.9/runpy.py # code object from '/usr/lib64/python3.9/__pycache__/runpy.cpython-39.pyc' # /usr/lib64/python3.9/importlib/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/importlib/__init__.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/warnings.cpython-39.pyc matches /usr/lib64/python3.9/warnings.py # code object from '/usr/lib64/python3.9/__pycache__/warnings.cpython-39.pyc' import 'warnings' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1041f40> import 'importlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1041e80> # /usr/lib64/python3.9/importlib/__pycache__/machinery.cpython-39.pyc matches /usr/lib64/python3.9/importlib/machinery.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/machinery.cpython-39.pyc' import 'importlib.machinery' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1041df0> # /usr/lib64/python3.9/importlib/__pycache__/util.cpython-39.pyc matches /usr/lib64/python3.9/importlib/util.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/util.cpython-39.pyc' # /usr/lib64/python3.9/importlib/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/importlib/abc.py # code object from '/usr/lib64/python3.9/importlib/__pycache__/abc.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/typing.cpython-39.pyc matches /usr/lib64/python3.9/typing.py # code object from '/usr/lib64/python3.9/__pycache__/typing.cpython-39.pyc' # /usr/lib64/python3.9/collections/__pycache__/abc.cpython-39.pyc matches /usr/lib64/python3.9/collections/abc.py # code object from '/usr/lib64/python3.9/collections/__pycache__/abc.cpython-39.pyc' import 'collections.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0dfa460> # /usr/lib64/python3.9/__pycache__/contextlib.cpython-39.pyc matches /usr/lib64/python3.9/contextlib.py # code object from '/usr/lib64/python3.9/__pycache__/contextlib.cpython-39.pyc' import 'contextlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0dfa550> import 'typing' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0dd80d0> import 'importlib.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1044b20> import 'importlib.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10444c0> # /usr/lib64/python3.9/__pycache__/pkgutil.cpython-39.pyc matches /usr/lib64/python3.9/pkgutil.py # code object from '/usr/lib64/python3.9/__pycache__/pkgutil.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/weakref.cpython-39.pyc matches /usr/lib64/python3.9/weakref.py # code object from '/usr/lib64/python3.9/__pycache__/weakref.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/_weakrefset.cpython-39.pyc matches /usr/lib64/python3.9/_weakrefset.py # code object from '/usr/lib64/python3.9/__pycache__/_weakrefset.cpython-39.pyc' import '_weakrefset' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0d142b0> import 'weakref' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0de5d60> import 'pkgutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1044fa0> import 'runpy' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10bb0d0> # /usr/lib64/python3.9/__pycache__/shutil.cpython-39.pyc matches /usr/lib64/python3.9/shutil.py # code object from '/usr/lib64/python3.9/__pycache__/shutil.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/fnmatch.cpython-39.pyc matches /usr/lib64/python3.9/fnmatch.py # code object from '/usr/lib64/python3.9/__pycache__/fnmatch.cpython-39.pyc' import 'fnmatch' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0d24be0> import 'errno' # # extension module 'zlib' loaded from '/usr/lib64/python3.9/lib-dynload/zlib.cpython-39-x86_64-linux-gnu.so' # extension module 'zlib' executed from '/usr/lib64/python3.9/lib-dynload/zlib.cpython-39-x86_64-linux-gnu.so' import 'zlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0d24f10> # /usr/lib64/python3.9/__pycache__/bz2.cpython-39.pyc matches /usr/lib64/python3.9/bz2.py # code object from '/usr/lib64/python3.9/__pycache__/bz2.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/_compression.cpython-39.pyc matches /usr/lib64/python3.9/_compression.py # code object from '/usr/lib64/python3.9/__pycache__/_compression.cpython-39.pyc' import '_compression' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0d37820> # /usr/lib64/python3.9/__pycache__/threading.cpython-39.pyc matches /usr/lib64/python3.9/threading.py # code object from '/usr/lib64/python3.9/__pycache__/threading.cpython-39.pyc' import 'threading' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0d37d60> # extension module '_bz2' loaded from '/usr/lib64/python3.9/lib-dynload/_bz2.cpython-39-x86_64-linux-gnu.so' # extension module '_bz2' executed from '/usr/lib64/python3.9/lib-dynload/_bz2.cpython-39-x86_64-linux-gnu.so' import '_bz2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cc5490> import 'bz2' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0d24f40> # /usr/lib64/python3.9/__pycache__/lzma.cpython-39.pyc matches /usr/lib64/python3.9/lzma.py # code object from '/usr/lib64/python3.9/__pycache__/lzma.cpython-39.pyc' # extension module '_lzma' loaded from '/usr/lib64/python3.9/lib-dynload/_lzma.cpython-39-x86_64-linux-gnu.so' # extension module '_lzma' executed from '/usr/lib64/python3.9/lib-dynload/_lzma.cpython-39-x86_64-linux-gnu.so' import '_lzma' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cd5370> import 'lzma' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0d376a0> import 'pwd' # # extension module 'grp' loaded from '/usr/lib64/python3.9/lib-dynload/grp.cpython-39-x86_64-linux-gnu.so' # extension module 'grp' executed from '/usr/lib64/python3.9/lib-dynload/grp.cpython-39-x86_64-linux-gnu.so' import 'grp' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cd5430> import 'shutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c1041ac0> # /usr/lib64/python3.9/__pycache__/tempfile.cpython-39.pyc matches /usr/lib64/python3.9/tempfile.py # code object from '/usr/lib64/python3.9/__pycache__/tempfile.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/random.cpython-39.pyc matches /usr/lib64/python3.9/random.py # code object from '/usr/lib64/python3.9/__pycache__/random.cpython-39.pyc' # extension module 'math' loaded from '/usr/lib64/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so' # extension module 'math' executed from '/usr/lib64/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so' import 'math' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cf1790> # /usr/lib64/python3.9/__pycache__/bisect.cpython-39.pyc matches /usr/lib64/python3.9/bisect.py # code object from '/usr/lib64/python3.9/__pycache__/bisect.cpython-39.pyc' # extension module '_bisect' loaded from '/usr/lib64/python3.9/lib-dynload/_bisect.cpython-39-x86_64-linux-gnu.so' # extension module '_bisect' executed from '/usr/lib64/python3.9/lib-dynload/_bisect.cpython-39-x86_64-linux-gnu.so' import '_bisect' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cf1a60> import 'bisect' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0cf1850> # extension module '_random' loaded from '/usr/lib64/python3.9/lib-dynload/_random.cpython-39-x86_64-linux-gnu.so' # extension module '_random' executed from '/usr/lib64/python3.9/lib-dynload/_random.cpython-39-x86_64-linux-gnu.so' import '_random' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cf1940> # /usr/lib64/python3.9/__pycache__/hashlib.cpython-39.pyc matches /usr/lib64/python3.9/hashlib.py # code object from '/usr/lib64/python3.9/__pycache__/hashlib.cpython-39.pyc' # extension module '_hashlib' loaded from '/usr/lib64/python3.9/lib-dynload/_hashlib.cpython-39-x86_64-linux-gnu.so' # extension module '_hashlib' executed from '/usr/lib64/python3.9/lib-dynload/_hashlib.cpython-39-x86_64-linux-gnu.so' import '_hashlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cf1d90> # extension module '_blake2' loaded from '/usr/lib64/python3.9/lib-dynload/_blake2.cpython-39-x86_64-linux-gnu.so' # extension module '_blake2' executed from '/usr/lib64/python3.9/lib-dynload/_blake2.cpython-39-x86_64-linux-gnu.so' import '_blake2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0cfb2e0> import 'hashlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0cf19d0> import 'random' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0ce5b20> import 'tempfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c10416a0> # /usr/lib64/python3.9/__pycache__/zipfile.cpython-39.pyc matches /usr/lib64/python3.9/zipfile.py # code object from '/usr/lib64/python3.9/__pycache__/zipfile.cpython-39.pyc' import 'zipfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0cf1b80> # code object from '/usr/lib64/python3.9/encodings/cp437.pyc' import 'encodings.cp437' # <_frozen_importlib_external.SourcelessFileLoader object at 0x7ff7c0c14760> # zipimport: found 30 names in '/tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip' # zipimport: zlib available # zipimport: zlib available import ansible # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/__init__.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/__future__.cpython-39.pyc matches /usr/lib64/python3.9/__future__.py # code object from '/usr/lib64/python3.9/__pycache__/__future__.cpython-39.pyc' import '__future__' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b3b8b0> # /usr/lib64/python3.9/json/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/json/__init__.py # code object from '/usr/lib64/python3.9/json/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/json/__pycache__/decoder.cpython-39.pyc matches /usr/lib64/python3.9/json/decoder.py # code object from '/usr/lib64/python3.9/json/__pycache__/decoder.cpython-39.pyc' # /usr/lib64/python3.9/json/__pycache__/scanner.cpython-39.pyc matches /usr/lib64/python3.9/json/scanner.py # code object from '/usr/lib64/python3.9/json/__pycache__/scanner.cpython-39.pyc' # extension module '_json' loaded from '/usr/lib64/python3.9/lib-dynload/_json.cpython-39-x86_64-linux-gnu.so' # extension module '_json' executed from '/usr/lib64/python3.9/lib-dynload/_json.cpython-39-x86_64-linux-gnu.so' import '_json' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0b3b160> import 'json.scanner' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b3b280> import 'json.decoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b3b5e0> # /usr/lib64/python3.9/json/__pycache__/encoder.cpython-39.pyc matches /usr/lib64/python3.9/json/encoder.py # code object from '/usr/lib64/python3.9/json/__pycache__/encoder.cpython-39.pyc' import 'json.encoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b3b4f0> import 'json' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b3be20> import 'atexit' # # extension module 'fcntl' loaded from '/usr/lib64/python3.9/lib-dynload/fcntl.cpython-39-x86_64-linux-gnu.so' # extension module 'fcntl' executed from '/usr/lib64/python3.9/lib-dynload/fcntl.cpython-39-x86_64-linux-gnu.so' import 'fcntl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0b3b580> # /usr/lib64/python3.9/__pycache__/locale.cpython-39.pyc matches /usr/lib64/python3.9/locale.py # code object from '/usr/lib64/python3.9/__pycache__/locale.cpython-39.pyc' import 'locale' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b3b100> # /usr/lib64/python3.9/__pycache__/platform.cpython-39.pyc matches /usr/lib64/python3.9/platform.py # code object from '/usr/lib64/python3.9/__pycache__/platform.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/subprocess.cpython-39.pyc matches /usr/lib64/python3.9/subprocess.py # code object from '/usr/lib64/python3.9/__pycache__/subprocess.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/signal.cpython-39.pyc matches /usr/lib64/python3.9/signal.py # code object from '/usr/lib64/python3.9/__pycache__/signal.cpython-39.pyc' import 'signal' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c056dfd0> # extension module '_posixsubprocess' loaded from '/usr/lib64/python3.9/lib-dynload/_posixsubprocess.cpython-39-x86_64-linux-gnu.so' # extension module '_posixsubprocess' executed from '/usr/lib64/python3.9/lib-dynload/_posixsubprocess.cpython-39-x86_64-linux-gnu.so' import '_posixsubprocess' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c058bc40> # extension module 'select' loaded from '/usr/lib64/python3.9/lib-dynload/select.cpython-39-x86_64-linux-gnu.so' # extension module 'select' executed from '/usr/lib64/python3.9/lib-dynload/select.cpython-39-x86_64-linux-gnu.so' import 'select' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c058bf40> # /usr/lib64/python3.9/__pycache__/selectors.cpython-39.pyc matches /usr/lib64/python3.9/selectors.py # code object from '/usr/lib64/python3.9/__pycache__/selectors.cpython-39.pyc' import 'selectors' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c058b2e0> import 'subprocess' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0ba3d90> import 'platform' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0ba33a0> # /usr/lib64/python3.9/__pycache__/shlex.cpython-39.pyc matches /usr/lib64/python3.9/shlex.py # code object from '/usr/lib64/python3.9/__pycache__/shlex.cpython-39.pyc' import 'shlex' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0ba3f40> # /usr/lib64/python3.9/__pycache__/traceback.cpython-39.pyc matches /usr/lib64/python3.9/traceback.py # code object from '/usr/lib64/python3.9/__pycache__/traceback.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/linecache.cpython-39.pyc matches /usr/lib64/python3.9/linecache.py # code object from '/usr/lib64/python3.9/__pycache__/linecache.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/tokenize.cpython-39.pyc matches /usr/lib64/python3.9/tokenize.py # code object from '/usr/lib64/python3.9/__pycache__/tokenize.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/token.cpython-39.pyc matches /usr/lib64/python3.9/token.py # code object from '/usr/lib64/python3.9/__pycache__/token.cpython-39.pyc' import 'token' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0c14a90> import 'tokenize' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05e9dc0> import 'linecache' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05e9490> import 'traceback' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b45580> # extension module 'syslog' loaded from '/usr/lib64/python3.9/lib-dynload/syslog.cpython-39-x86_64-linux-gnu.so' # extension module 'syslog' executed from '/usr/lib64/python3.9/lib-dynload/syslog.cpython-39-x86_64-linux-gnu.so' import 'syslog' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c05e95b0> # /usr/lib64/python3.9/site-packages/systemd/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/__init__.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/__init__.cpython-39.pyc' import 'systemd' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05e95e0> # /usr/lib64/python3.9/site-packages/systemd/__pycache__/journal.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/journal.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/journal.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/datetime.cpython-39.pyc matches /usr/lib64/python3.9/datetime.py # code object from '/usr/lib64/python3.9/__pycache__/datetime.cpython-39.pyc' # extension module '_datetime' loaded from '/usr/lib64/python3.9/lib-dynload/_datetime.cpython-39-x86_64-linux-gnu.so' # extension module '_datetime' executed from '/usr/lib64/python3.9/lib-dynload/_datetime.cpython-39-x86_64-linux-gnu.so' import '_datetime' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c055ef70> import 'datetime' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b832e0> # /usr/lib64/python3.9/__pycache__/uuid.cpython-39.pyc matches /usr/lib64/python3.9/uuid.py # code object from '/usr/lib64/python3.9/__pycache__/uuid.cpython-39.pyc' # extension module '_uuid' loaded from '/usr/lib64/python3.9/lib-dynload/_uuid.cpython-39-x86_64-linux-gnu.so' # extension module '_uuid' executed from '/usr/lib64/python3.9/lib-dynload/_uuid.cpython-39-x86_64-linux-gnu.so' import '_uuid' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c055b7f0> import 'uuid' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b83460> # /usr/lib64/python3.9/logging/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/logging/__init__.py # code object from '/usr/lib64/python3.9/logging/__pycache__/__init__.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/string.cpython-39.pyc matches /usr/lib64/python3.9/string.py # code object from '/usr/lib64/python3.9/__pycache__/string.cpython-39.pyc' import '_string' # import 'string' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b9bf40> import 'logging' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c055b790> # extension module 'systemd._journal' loaded from '/usr/lib64/python3.9/site-packages/systemd/_journal.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._journal' executed from '/usr/lib64/python3.9/site-packages/systemd/_journal.cpython-39-x86_64-linux-gnu.so' import 'systemd._journal' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c055b5e0> # extension module 'systemd._reader' loaded from '/usr/lib64/python3.9/site-packages/systemd/_reader.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._reader' executed from '/usr/lib64/python3.9/site-packages/systemd/_reader.cpython-39-x86_64-linux-gnu.so' import 'systemd._reader' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c055a550> # extension module 'systemd.id128' loaded from '/usr/lib64/python3.9/site-packages/systemd/id128.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd.id128' executed from '/usr/lib64/python3.9/site-packages/systemd/id128.cpython-39-x86_64-linux-gnu.so' import 'systemd.id128' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c055a490> import 'systemd.journal' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b7c9a0> # /usr/lib64/python3.9/site-packages/systemd/__pycache__/daemon.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/systemd/daemon.py # code object from '/usr/lib64/python3.9/site-packages/systemd/__pycache__/daemon.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/socket.cpython-39.pyc matches /usr/lib64/python3.9/socket.py # code object from '/usr/lib64/python3.9/__pycache__/socket.cpython-39.pyc' # extension module '_socket' loaded from '/usr/lib64/python3.9/lib-dynload/_socket.cpython-39-x86_64-linux-gnu.so' # extension module '_socket' executed from '/usr/lib64/python3.9/lib-dynload/_socket.cpython-39-x86_64-linux-gnu.so' import '_socket' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c05df6a0> # extension module 'array' loaded from '/usr/lib64/python3.9/lib-dynload/array.cpython-39-x86_64-linux-gnu.so' # extension module 'array' executed from '/usr/lib64/python3.9/lib-dynload/array.cpython-39-x86_64-linux-gnu.so' import 'array' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c05debb0> import 'socket' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05ef0d0> # extension module 'systemd._daemon' loaded from '/usr/lib64/python3.9/site-packages/systemd/_daemon.cpython-39-x86_64-linux-gnu.so' # extension module 'systemd._daemon' executed from '/usr/lib64/python3.9/site-packages/systemd/_daemon.cpython-39-x86_64-linux-gnu.so' import 'systemd._daemon' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c05df100> import 'systemd.daemon' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c0b47c40> # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.compat # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/compat/__init__.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.text # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/text/__init__.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.six # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/six/__init__.py import 'ansible.module_utils.six.moves' # import 'ansible.module_utils.six.moves.collections_abc' # import ansible.module_utils.common.text.converters # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/text/converters.py # /usr/lib64/python3.9/ctypes/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/ctypes/__init__.py # code object from '/usr/lib64/python3.9/ctypes/__pycache__/__init__.cpython-39.pyc' # extension module '_ctypes' loaded from '/usr/lib64/python3.9/lib-dynload/_ctypes.cpython-39-x86_64-linux-gnu.so' # extension module '_ctypes' executed from '/usr/lib64/python3.9/lib-dynload/_ctypes.cpython-39-x86_64-linux-gnu.so' import '_ctypes' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0526940> # /usr/lib64/python3.9/ctypes/__pycache__/_endian.cpython-39.pyc matches /usr/lib64/python3.9/ctypes/_endian.py # code object from '/usr/lib64/python3.9/ctypes/__pycache__/_endian.cpython-39.pyc' import 'ctypes._endian' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05dcd30> import 'ctypes' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05d37c0> import ansible.module_utils.compat.selinux # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/compat/selinux.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils._text # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/_text.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/copy.cpython-39.pyc matches /usr/lib64/python3.9/copy.py # code object from '/usr/lib64/python3.9/__pycache__/copy.cpython-39.pyc' import 'copy' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c05de4c0> # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.collections # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/collections.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.warnings # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/warnings.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.errors # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/errors.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.parsing # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/parsing/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.parsing.convert_bool # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/parsing/convert_bool.py # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/__pycache__/ast.cpython-39.pyc matches /usr/lib64/python3.9/ast.py # code object from '/usr/lib64/python3.9/__pycache__/ast.cpython-39.pyc' import '_ast' # import 'ast' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c00be940> # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.text.formatters # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/text/formatters.py import ansible.module_utils.common.validation # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/validation.py import ansible.module_utils.common.parameters # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/parameters.py import ansible.module_utils.common.arg_spec # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/arg_spec.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common.locale # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/locale.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.9/site-packages/selinux/__pycache__/__init__.cpython-39.pyc matches /usr/lib64/python3.9/site-packages/selinux/__init__.py # code object from '/usr/lib64/python3.9/site-packages/selinux/__pycache__/__init__.cpython-39.pyc' # extension module 'selinux._selinux' loaded from '/usr/lib64/python3.9/site-packages/selinux/_selinux.cpython-39-x86_64-linux-gnu.so' # extension module 'selinux._selinux' executed from '/usr/lib64/python3.9/site-packages/selinux/_selinux.cpython-39-x86_64-linux-gnu.so' import 'selinux._selinux' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7ff7c0b8eb50> import 'selinux' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c00bd070> import ansible.module_utils.common.file # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/file.py import ansible.module_utils.common.process # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/process.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib/python3.9/site-packages/__pycache__/distro.cpython-39.pyc matches /usr/lib/python3.9/site-packages/distro.py # code object from '/usr/lib/python3.9/site-packages/__pycache__/distro.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/argparse.cpython-39.pyc matches /usr/lib64/python3.9/argparse.py # code object from '/usr/lib64/python3.9/__pycache__/argparse.cpython-39.pyc' # /usr/lib64/python3.9/__pycache__/gettext.cpython-39.pyc matches /usr/lib64/python3.9/gettext.py # code object from '/usr/lib64/python3.9/__pycache__/gettext.cpython-39.pyc' import 'gettext' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c010e6d0> import 'argparse' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c051dc10> import 'distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7ff7c051c5b0> # destroy ansible.module_utils.distro import ansible.module_utils.distro # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/distro/__init__.py # zipimport: zlib available # zipimport: zlib available import ansible.module_utils.common._utils # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/_utils.py import ansible.module_utils.common.sys_info # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/common/sys_info.py import ansible.module_utils.basic # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/module_utils/basic.py # zipimport: zlib available # zipimport: zlib available import ansible.modules # loaded from Zip /tmp/ansible_stat_payload_qgrccisv/ansible_stat_payload.zip/ansible/modules/__init__.py # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/run/ostree-booted", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} # destroy __main__ # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.path_hooks # clear sys.path_importer_cache # clear sys.meta_path # clear sys.__interactivehook__ # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing encodings.latin_1 # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing _locale # cleanup[2] removing _bootlocale # destroy _bootlocale # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing sre_constants # destroy sre_constants # cleanup[2] removing sre_parse # cleanup[2] removing sre_compile # cleanup[2] removing _heapq # cleanup[2] removing heapq # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # destroy base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing collections.abc # cleanup[2] removing contextlib # cleanup[2] removing typing # destroy typing # cleanup[2] removing importlib.abc # cleanup[2] removing importlib.util # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing threading # cleanup[2] removing _bz2 # destroy _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing pwd # cleanup[2] removing grp # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing tempfile # cleanup[2] removing zipfile # destroy zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing fcntl # cleanup[2] removing locale # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing subprocess # cleanup[2] removing platform # cleanup[2] removing shlex # cleanup[2] removing token # destroy token # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # destroy socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # destroy _sitebuiltins # destroy importlib.util # destroy importlib.abc # destroy importlib.machinery # destroy zipimport # destroy _compression # destroy binascii # destroy importlib # destroy struct # destroy bz2 # destroy lzma # destroy __main__ # destroy locale # destroy tempfile # destroy systemd.journal # destroy systemd.daemon # destroy ansible.module_utils.compat.selinux # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy encodings # destroy syslog # destroy uuid # destroy array # destroy datetime # destroy selinux # destroy distro # destroy json # destroy shlex # destroy logging # destroy argparse # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # cleanup[3] wiping tokenize # cleanup[3] wiping platform # destroy subprocess # cleanup[3] wiping selectors # cleanup[3] wiping select # cleanup[3] wiping _posixsubprocess # cleanup[3] wiping signal # cleanup[3] wiping fcntl # cleanup[3] wiping atexit # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping _blake2 # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping shutil # destroy fnmatch # cleanup[3] wiping grp # cleanup[3] wiping pwd # cleanup[3] wiping _lzma # cleanup[3] wiping threading # cleanup[3] wiping zlib # cleanup[3] wiping errno # cleanup[3] wiping weakref # cleanup[3] wiping contextlib # cleanup[3] wiping collections.abc # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy enum # destroy sre_compile # destroy copyreg # cleanup[3] wiping functools # cleanup[3] wiping _functools # destroy _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy heapq # destroy collections.abc # cleanup[3] wiping _collections # destroy _collections # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping itertools # cleanup[3] wiping _heapq # cleanup[3] wiping sre_parse # cleanup[3] wiping _sre # cleanup[3] wiping types # cleanup[3] wiping _locale # destroy _locale # cleanup[3] wiping os # cleanup[3] wiping os.path # destroy genericpath # cleanup[3] wiping posixpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.latin_1 # cleanup[3] wiping _signal # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy systemd._daemon # destroy _socket # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy fcntl # destroy _blake2 # destroy _lzma # destroy zlib # destroy _signal # destroy platform # destroy _uuid # destroy _sre # destroy sre_parse # destroy tokenize # destroy _heapq # destroy posixpath # destroy stat # destroy ansible.module_utils.six.moves.urllib # destroy errno # destroy signal # destroy contextlib # destroy pwd # destroy grp # destroy _posixsubprocess # destroy selectors # destroy select # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy itertools # destroy operator # destroy ansible.module_utils.six.moves # destroy _operator # destroy _frozen_importlib_external # destroy _imp # destroy io # destroy marshal # destroy _frozen_importlib # clear sys.audit hooks , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. [WARNING]: Module invocation had junk after the JSON data: # destroy __main__ # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.path_hooks # clear sys.path_importer_cache # clear sys.meta_path # clear sys.__interactivehook__ # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing encodings.latin_1 # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing _locale # cleanup[2] removing _bootlocale # destroy _bootlocale # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing sre_constants # destroy sre_constants # cleanup[2] removing sre_parse # cleanup[2] removing sre_compile # cleanup[2] removing _heapq # cleanup[2] removing heapq # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # destroy base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing collections.abc # cleanup[2] removing contextlib # cleanup[2] removing typing # destroy typing # cleanup[2] removing importlib.abc # cleanup[2] removing importlib.util # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing threading # cleanup[2] removing _bz2 # destroy _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing pwd # cleanup[2] removing grp # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing tempfile # cleanup[2] removing zipfile # destroy zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing fcntl # cleanup[2] removing locale # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing subprocess # cleanup[2] removing platform # cleanup[2] removing shlex # cleanup[2] removing token # destroy token # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # destroy socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # destroy _sitebuiltins # destroy importlib.util # destroy importlib.abc # destroy importlib.machinery # destroy zipimport # destroy _compression # destroy binascii # destroy importlib # destroy struct # destroy bz2 # destroy lzma # destroy __main__ # destroy locale # destroy tempfile # destroy systemd.journal # destroy systemd.daemon # destroy ansible.module_utils.compat.selinux # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy encodings # destroy syslog # destroy uuid # destroy array # destroy datetime # destroy selinux # destroy distro # destroy json # destroy shlex # destroy logging # destroy argparse # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # cleanup[3] wiping tokenize # cleanup[3] wiping platform # destroy subprocess # cleanup[3] wiping selectors # cleanup[3] wiping select # cleanup[3] wiping _posixsubprocess # cleanup[3] wiping signal # cleanup[3] wiping fcntl # cleanup[3] wiping atexit # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping _blake2 # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping shutil # destroy fnmatch # cleanup[3] wiping grp # cleanup[3] wiping pwd # cleanup[3] wiping _lzma # cleanup[3] wiping threading # cleanup[3] wiping zlib # cleanup[3] wiping errno # cleanup[3] wiping weakref # cleanup[3] wiping contextlib # cleanup[3] wiping collections.abc # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy enum # destroy sre_compile # destroy copyreg # cleanup[3] wiping functools # cleanup[3] wiping _functools # destroy _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy heapq # destroy collections.abc # cleanup[3] wiping _collections # destroy _collections # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping itertools # cleanup[3] wiping _heapq # cleanup[3] wiping sre_parse # cleanup[3] wiping _sre # cleanup[3] wiping types # cleanup[3] wiping _locale # destroy _locale # cleanup[3] wiping os # cleanup[3] wiping os.path # destroy genericpath # cleanup[3] wiping posixpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.latin_1 # cleanup[3] wiping _signal # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy systemd._daemon # destroy _socket # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy fcntl # destroy _blake2 # destroy _lzma # destroy zlib # destroy _signal # destroy platform # destroy _uuid # destroy _sre # destroy sre_parse # destroy tokenize # destroy _heapq # destroy posixpath # destroy stat # destroy ansible.module_utils.six.moves.urllib # destroy errno # destroy signal # destroy contextlib # destroy pwd # destroy grp # destroy _posixsubprocess # destroy selectors # destroy select # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy itertools # destroy operator # destroy ansible.module_utils.six.moves # destroy _operator # destroy _frozen_importlib_external # destroy _imp # destroy io # destroy marshal # destroy _frozen_importlib # clear sys.audit hooks 13118 1727204065.74751: done with _execute_module (stat, {'path': '/run/ostree-booted', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204065.1472027-13731-146327907219557/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204065.74755: _low_level_execute_command(): starting 13118 1727204065.74758: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204065.1472027-13731-146327907219557/ > /dev/null 2>&1 && sleep 0' 13118 1727204065.74884: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204065.74888: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204065.74933: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.74936: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204065.74939: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204065.75021: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204065.75024: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204065.75031: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204065.75085: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204065.77639: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204065.77643: stderr chunk (state=3): >>><<< 13118 1727204065.77646: stdout chunk (state=3): >>><<< 13118 1727204065.77746: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204065.77750: handler run complete 13118 1727204065.77753: attempt loop complete, returning result 13118 1727204065.77755: _execute() done 13118 1727204065.77757: dumping result to json 13118 1727204065.77759: done dumping result, returning 13118 1727204065.77761: done running TaskExecutor() for managed-node2/TASK: Check if system is ostree [0affcd87-79f5-56a3-0a64-0000000000df] 13118 1727204065.77765: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000df 13118 1727204065.77824: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000df 13118 1727204065.77830: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "stat": { "exists": false } } 13118 1727204065.77906: no more pending results, returning what we have 13118 1727204065.77909: results queue empty 13118 1727204065.77910: checking for any_errors_fatal 13118 1727204065.77919: done checking for any_errors_fatal 13118 1727204065.77920: checking for max_fail_percentage 13118 1727204065.77922: done checking for max_fail_percentage 13118 1727204065.77923: checking to see if all hosts have failed and the running result is not ok 13118 1727204065.77923: done checking to see if all hosts have failed 13118 1727204065.77924: getting the remaining hosts for this loop 13118 1727204065.77925: done getting the remaining hosts for this loop 13118 1727204065.77932: getting the next task for host managed-node2 13118 1727204065.77938: done getting next task for host managed-node2 13118 1727204065.77940: ^ task is: TASK: Set flag to indicate system is ostree 13118 1727204065.77943: ^ state is: HOST STATE: block=2, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204065.77947: getting variables 13118 1727204065.77948: in VariableManager get_vars() 13118 1727204065.77979: Calling all_inventory to load vars for managed-node2 13118 1727204065.77981: Calling groups_inventory to load vars for managed-node2 13118 1727204065.77985: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204065.77996: Calling all_plugins_play to load vars for managed-node2 13118 1727204065.77999: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204065.78001: Calling groups_plugins_play to load vars for managed-node2 13118 1727204065.78156: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204065.78325: done with get_vars() 13118 1727204065.78337: done getting variables 13118 1727204065.78424: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=False, class_only=True) TASK [Set flag to indicate system is ostree] *********************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:22 Tuesday 24 September 2024 14:54:25 -0400 (0:00:00.737) 0:00:03.084 ***** 13118 1727204065.78452: entering _queue_task() for managed-node2/set_fact 13118 1727204065.78453: Creating lock for set_fact 13118 1727204065.78741: worker is 1 (out of 1 available) 13118 1727204065.78754: exiting _queue_task() for managed-node2/set_fact 13118 1727204065.78768: done queuing things up, now waiting for results queue to drain 13118 1727204065.78769: waiting for pending results... 13118 1727204065.78922: running TaskExecutor() for managed-node2/TASK: Set flag to indicate system is ostree 13118 1727204065.79001: in run() - task 0affcd87-79f5-56a3-0a64-0000000000e0 13118 1727204065.79011: variable 'ansible_search_path' from source: unknown 13118 1727204065.79015: variable 'ansible_search_path' from source: unknown 13118 1727204065.79045: calling self._execute() 13118 1727204065.79100: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.79110: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.79118: variable 'omit' from source: magic vars 13118 1727204065.79452: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204065.79627: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204065.79662: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204065.79692: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204065.79722: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204065.79791: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204065.79836: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204065.79840: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204065.79859: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204065.79951: Evaluated conditional (not __network_is_ostree is defined): True 13118 1727204065.79957: variable 'omit' from source: magic vars 13118 1727204065.79986: variable 'omit' from source: magic vars 13118 1727204065.80071: variable '__ostree_booted_stat' from source: set_fact 13118 1727204065.80109: variable 'omit' from source: magic vars 13118 1727204065.80130: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204065.80149: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204065.80362: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204065.80389: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204065.80404: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204065.80436: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204065.80445: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.80453: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.80550: Set connection var ansible_timeout to 10 13118 1727204065.80570: Set connection var ansible_pipelining to False 13118 1727204065.80578: Set connection var ansible_connection to ssh 13118 1727204065.80589: Set connection var ansible_shell_executable to /bin/sh 13118 1727204065.80598: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204065.80604: Set connection var ansible_shell_type to sh 13118 1727204065.80631: variable 'ansible_shell_executable' from source: unknown 13118 1727204065.80639: variable 'ansible_connection' from source: unknown 13118 1727204065.80646: variable 'ansible_module_compression' from source: unknown 13118 1727204065.80652: variable 'ansible_shell_type' from source: unknown 13118 1727204065.80659: variable 'ansible_shell_executable' from source: unknown 13118 1727204065.80667: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.80675: variable 'ansible_pipelining' from source: unknown 13118 1727204065.80681: variable 'ansible_timeout' from source: unknown 13118 1727204065.80689: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.80791: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204065.80808: variable 'omit' from source: magic vars 13118 1727204065.80819: starting attempt loop 13118 1727204065.80827: running the handler 13118 1727204065.80846: handler run complete 13118 1727204065.80858: attempt loop complete, returning result 13118 1727204065.80866: _execute() done 13118 1727204065.80872: dumping result to json 13118 1727204065.80879: done dumping result, returning 13118 1727204065.80888: done running TaskExecutor() for managed-node2/TASK: Set flag to indicate system is ostree [0affcd87-79f5-56a3-0a64-0000000000e0] 13118 1727204065.80897: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000e0 ok: [managed-node2] => { "ansible_facts": { "__network_is_ostree": false }, "changed": false } 13118 1727204065.81041: no more pending results, returning what we have 13118 1727204065.81044: results queue empty 13118 1727204065.81045: checking for any_errors_fatal 13118 1727204065.81051: done checking for any_errors_fatal 13118 1727204065.81051: checking for max_fail_percentage 13118 1727204065.81053: done checking for max_fail_percentage 13118 1727204065.81054: checking to see if all hosts have failed and the running result is not ok 13118 1727204065.81054: done checking to see if all hosts have failed 13118 1727204065.81055: getting the remaining hosts for this loop 13118 1727204065.81056: done getting the remaining hosts for this loop 13118 1727204065.81060: getting the next task for host managed-node2 13118 1727204065.81075: done getting next task for host managed-node2 13118 1727204065.81078: ^ task is: TASK: Fix CentOS6 Base repo 13118 1727204065.81080: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204065.81083: getting variables 13118 1727204065.81085: in VariableManager get_vars() 13118 1727204065.81113: Calling all_inventory to load vars for managed-node2 13118 1727204065.81115: Calling groups_inventory to load vars for managed-node2 13118 1727204065.81118: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204065.81132: Calling all_plugins_play to load vars for managed-node2 13118 1727204065.81134: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204065.81138: Calling groups_plugins_play to load vars for managed-node2 13118 1727204065.81308: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000e0 13118 1727204065.81330: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204065.81545: done with get_vars() 13118 1727204065.81556: done getting variables 13118 1727204065.81594: WORKER PROCESS EXITING 13118 1727204065.81697: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=False, class_only=True) TASK [Fix CentOS6 Base repo] *************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:26 Tuesday 24 September 2024 14:54:25 -0400 (0:00:00.032) 0:00:03.117 ***** 13118 1727204065.81735: entering _queue_task() for managed-node2/copy 13118 1727204065.82043: worker is 1 (out of 1 available) 13118 1727204065.82071: exiting _queue_task() for managed-node2/copy 13118 1727204065.82084: done queuing things up, now waiting for results queue to drain 13118 1727204065.82086: waiting for pending results... 13118 1727204065.82298: running TaskExecutor() for managed-node2/TASK: Fix CentOS6 Base repo 13118 1727204065.82362: in run() - task 0affcd87-79f5-56a3-0a64-0000000000e2 13118 1727204065.82381: variable 'ansible_search_path' from source: unknown 13118 1727204065.82385: variable 'ansible_search_path' from source: unknown 13118 1727204065.82410: calling self._execute() 13118 1727204065.82465: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.82469: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.82480: variable 'omit' from source: magic vars 13118 1727204065.82805: variable 'ansible_distribution' from source: facts 13118 1727204065.82822: Evaluated conditional (ansible_distribution == 'CentOS'): True 13118 1727204065.82903: variable 'ansible_distribution_major_version' from source: facts 13118 1727204065.82907: Evaluated conditional (ansible_distribution_major_version == '6'): False 13118 1727204065.82910: when evaluation is False, skipping this task 13118 1727204065.82914: _execute() done 13118 1727204065.82919: dumping result to json 13118 1727204065.82921: done dumping result, returning 13118 1727204065.82926: done running TaskExecutor() for managed-node2/TASK: Fix CentOS6 Base repo [0affcd87-79f5-56a3-0a64-0000000000e2] 13118 1727204065.82940: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000e2 13118 1727204065.83031: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000e2 13118 1727204065.83035: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_distribution_major_version == '6'", "skip_reason": "Conditional result was False" } 13118 1727204065.83122: no more pending results, returning what we have 13118 1727204065.83124: results queue empty 13118 1727204065.83125: checking for any_errors_fatal 13118 1727204065.83131: done checking for any_errors_fatal 13118 1727204065.83132: checking for max_fail_percentage 13118 1727204065.83134: done checking for max_fail_percentage 13118 1727204065.83134: checking to see if all hosts have failed and the running result is not ok 13118 1727204065.83135: done checking to see if all hosts have failed 13118 1727204065.83136: getting the remaining hosts for this loop 13118 1727204065.83137: done getting the remaining hosts for this loop 13118 1727204065.83140: getting the next task for host managed-node2 13118 1727204065.83145: done getting next task for host managed-node2 13118 1727204065.83147: ^ task is: TASK: Include the task 'enable_epel.yml' 13118 1727204065.83150: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204065.83153: getting variables 13118 1727204065.83154: in VariableManager get_vars() 13118 1727204065.83179: Calling all_inventory to load vars for managed-node2 13118 1727204065.83182: Calling groups_inventory to load vars for managed-node2 13118 1727204065.83184: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204065.83191: Calling all_plugins_play to load vars for managed-node2 13118 1727204065.83193: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204065.83195: Calling groups_plugins_play to load vars for managed-node2 13118 1727204065.83444: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204065.83551: done with get_vars() 13118 1727204065.83557: done getting variables TASK [Include the task 'enable_epel.yml'] ************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:51 Tuesday 24 September 2024 14:54:25 -0400 (0:00:00.018) 0:00:03.136 ***** 13118 1727204065.83621: entering _queue_task() for managed-node2/include_tasks 13118 1727204065.83812: worker is 1 (out of 1 available) 13118 1727204065.83826: exiting _queue_task() for managed-node2/include_tasks 13118 1727204065.83839: done queuing things up, now waiting for results queue to drain 13118 1727204065.83840: waiting for pending results... 13118 1727204065.84041: running TaskExecutor() for managed-node2/TASK: Include the task 'enable_epel.yml' 13118 1727204065.84151: in run() - task 0affcd87-79f5-56a3-0a64-0000000000e3 13118 1727204065.84172: variable 'ansible_search_path' from source: unknown 13118 1727204065.84179: variable 'ansible_search_path' from source: unknown 13118 1727204065.84215: calling self._execute() 13118 1727204065.84300: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.84310: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.84323: variable 'omit' from source: magic vars 13118 1727204065.84831: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204065.87285: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204065.87354: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204065.87435: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204065.87474: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204065.87505: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204065.87582: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204065.87626: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204065.87670: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204065.87730: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204065.87748: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204065.87861: variable '__network_is_ostree' from source: set_fact 13118 1727204065.87885: Evaluated conditional (not __network_is_ostree | d(false)): True 13118 1727204065.87894: _execute() done 13118 1727204065.87900: dumping result to json 13118 1727204065.87905: done dumping result, returning 13118 1727204065.87913: done running TaskExecutor() for managed-node2/TASK: Include the task 'enable_epel.yml' [0affcd87-79f5-56a3-0a64-0000000000e3] 13118 1727204065.87920: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000e3 13118 1727204065.88025: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000e3 13118 1727204065.88033: WORKER PROCESS EXITING 13118 1727204065.88059: no more pending results, returning what we have 13118 1727204065.88066: in VariableManager get_vars() 13118 1727204065.88100: Calling all_inventory to load vars for managed-node2 13118 1727204065.88103: Calling groups_inventory to load vars for managed-node2 13118 1727204065.88106: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204065.88117: Calling all_plugins_play to load vars for managed-node2 13118 1727204065.88120: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204065.88122: Calling groups_plugins_play to load vars for managed-node2 13118 1727204065.88288: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204065.88462: done with get_vars() 13118 1727204065.88470: variable 'ansible_search_path' from source: unknown 13118 1727204065.88471: variable 'ansible_search_path' from source: unknown 13118 1727204065.88497: we have included files to process 13118 1727204065.88498: generating all_blocks data 13118 1727204065.88499: done generating all_blocks data 13118 1727204065.88502: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml 13118 1727204065.88503: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml 13118 1727204065.88505: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml 13118 1727204065.88973: done processing included file 13118 1727204065.88975: iterating over new_blocks loaded from include file 13118 1727204065.88976: in VariableManager get_vars() 13118 1727204065.88985: done with get_vars() 13118 1727204065.88986: filtering new block on tags 13118 1727204065.89000: done filtering new block on tags 13118 1727204065.89002: in VariableManager get_vars() 13118 1727204065.89008: done with get_vars() 13118 1727204065.89009: filtering new block on tags 13118 1727204065.89016: done filtering new block on tags 13118 1727204065.89017: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml for managed-node2 13118 1727204065.89021: extending task lists for all hosts with included blocks 13118 1727204065.89086: done extending task lists 13118 1727204065.89087: done processing included files 13118 1727204065.89088: results queue empty 13118 1727204065.89088: checking for any_errors_fatal 13118 1727204065.89091: done checking for any_errors_fatal 13118 1727204065.89091: checking for max_fail_percentage 13118 1727204065.89092: done checking for max_fail_percentage 13118 1727204065.89092: checking to see if all hosts have failed and the running result is not ok 13118 1727204065.89093: done checking to see if all hosts have failed 13118 1727204065.89093: getting the remaining hosts for this loop 13118 1727204065.89094: done getting the remaining hosts for this loop 13118 1727204065.89095: getting the next task for host managed-node2 13118 1727204065.89098: done getting next task for host managed-node2 13118 1727204065.89100: ^ task is: TASK: Create EPEL {{ ansible_distribution_major_version }} 13118 1727204065.89102: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204065.89103: getting variables 13118 1727204065.89104: in VariableManager get_vars() 13118 1727204065.89109: Calling all_inventory to load vars for managed-node2 13118 1727204065.89111: Calling groups_inventory to load vars for managed-node2 13118 1727204065.89112: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204065.89116: Calling all_plugins_play to load vars for managed-node2 13118 1727204065.89122: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204065.89123: Calling groups_plugins_play to load vars for managed-node2 13118 1727204065.89207: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204065.89325: done with get_vars() 13118 1727204065.89332: done getting variables 13118 1727204065.89380: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=False, class_only=True) 13118 1727204065.89548: variable 'ansible_distribution_major_version' from source: facts TASK [Create EPEL 9] *********************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:8 Tuesday 24 September 2024 14:54:25 -0400 (0:00:00.059) 0:00:03.195 ***** 13118 1727204065.89593: entering _queue_task() for managed-node2/command 13118 1727204065.89594: Creating lock for command 13118 1727204065.89907: worker is 1 (out of 1 available) 13118 1727204065.89921: exiting _queue_task() for managed-node2/command 13118 1727204065.89934: done queuing things up, now waiting for results queue to drain 13118 1727204065.89936: waiting for pending results... 13118 1727204065.90205: running TaskExecutor() for managed-node2/TASK: Create EPEL 9 13118 1727204065.90333: in run() - task 0affcd87-79f5-56a3-0a64-0000000000fd 13118 1727204065.90360: variable 'ansible_search_path' from source: unknown 13118 1727204065.90372: variable 'ansible_search_path' from source: unknown 13118 1727204065.90416: calling self._execute() 13118 1727204065.90512: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.90524: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.90535: variable 'omit' from source: magic vars 13118 1727204065.90821: variable 'ansible_distribution' from source: facts 13118 1727204065.90829: Evaluated conditional (ansible_distribution in ['RedHat', 'CentOS']): True 13118 1727204065.90923: variable 'ansible_distribution_major_version' from source: facts 13118 1727204065.90926: Evaluated conditional (ansible_distribution_major_version in ['7', '8']): False 13118 1727204065.90929: when evaluation is False, skipping this task 13118 1727204065.90932: _execute() done 13118 1727204065.90937: dumping result to json 13118 1727204065.90941: done dumping result, returning 13118 1727204065.90946: done running TaskExecutor() for managed-node2/TASK: Create EPEL 9 [0affcd87-79f5-56a3-0a64-0000000000fd] 13118 1727204065.90952: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000fd 13118 1727204065.91048: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000fd 13118 1727204065.91050: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_distribution_major_version in ['7', '8']", "skip_reason": "Conditional result was False" } 13118 1727204065.91104: no more pending results, returning what we have 13118 1727204065.91108: results queue empty 13118 1727204065.91109: checking for any_errors_fatal 13118 1727204065.91110: done checking for any_errors_fatal 13118 1727204065.91111: checking for max_fail_percentage 13118 1727204065.91113: done checking for max_fail_percentage 13118 1727204065.91113: checking to see if all hosts have failed and the running result is not ok 13118 1727204065.91114: done checking to see if all hosts have failed 13118 1727204065.91115: getting the remaining hosts for this loop 13118 1727204065.91116: done getting the remaining hosts for this loop 13118 1727204065.91119: getting the next task for host managed-node2 13118 1727204065.91125: done getting next task for host managed-node2 13118 1727204065.91127: ^ task is: TASK: Install yum-utils package 13118 1727204065.91130: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204065.91133: getting variables 13118 1727204065.91135: in VariableManager get_vars() 13118 1727204065.91161: Calling all_inventory to load vars for managed-node2 13118 1727204065.91163: Calling groups_inventory to load vars for managed-node2 13118 1727204065.91166: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204065.91177: Calling all_plugins_play to load vars for managed-node2 13118 1727204065.91179: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204065.91181: Calling groups_plugins_play to load vars for managed-node2 13118 1727204065.91293: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204065.91406: done with get_vars() 13118 1727204065.91413: done getting variables 13118 1727204065.91485: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=False, class_only=True) TASK [Install yum-utils package] *********************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:26 Tuesday 24 September 2024 14:54:25 -0400 (0:00:00.019) 0:00:03.215 ***** 13118 1727204065.91505: entering _queue_task() for managed-node2/package 13118 1727204065.91507: Creating lock for package 13118 1727204065.91700: worker is 1 (out of 1 available) 13118 1727204065.91713: exiting _queue_task() for managed-node2/package 13118 1727204065.91726: done queuing things up, now waiting for results queue to drain 13118 1727204065.91727: waiting for pending results... 13118 1727204065.91873: running TaskExecutor() for managed-node2/TASK: Install yum-utils package 13118 1727204065.91939: in run() - task 0affcd87-79f5-56a3-0a64-0000000000fe 13118 1727204065.91949: variable 'ansible_search_path' from source: unknown 13118 1727204065.91952: variable 'ansible_search_path' from source: unknown 13118 1727204065.91985: calling self._execute() 13118 1727204065.92038: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.92041: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.92050: variable 'omit' from source: magic vars 13118 1727204065.92358: variable 'ansible_distribution' from source: facts 13118 1727204065.92370: Evaluated conditional (ansible_distribution in ['RedHat', 'CentOS']): True 13118 1727204065.92457: variable 'ansible_distribution_major_version' from source: facts 13118 1727204065.92461: Evaluated conditional (ansible_distribution_major_version in ['7', '8']): False 13118 1727204065.92466: when evaluation is False, skipping this task 13118 1727204065.92469: _execute() done 13118 1727204065.92471: dumping result to json 13118 1727204065.92474: done dumping result, returning 13118 1727204065.92480: done running TaskExecutor() for managed-node2/TASK: Install yum-utils package [0affcd87-79f5-56a3-0a64-0000000000fe] 13118 1727204065.92486: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000fe 13118 1727204065.92582: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000fe 13118 1727204065.92585: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_distribution_major_version in ['7', '8']", "skip_reason": "Conditional result was False" } 13118 1727204065.92648: no more pending results, returning what we have 13118 1727204065.92650: results queue empty 13118 1727204065.92651: checking for any_errors_fatal 13118 1727204065.92655: done checking for any_errors_fatal 13118 1727204065.92656: checking for max_fail_percentage 13118 1727204065.92657: done checking for max_fail_percentage 13118 1727204065.92658: checking to see if all hosts have failed and the running result is not ok 13118 1727204065.92658: done checking to see if all hosts have failed 13118 1727204065.92659: getting the remaining hosts for this loop 13118 1727204065.92660: done getting the remaining hosts for this loop 13118 1727204065.92665: getting the next task for host managed-node2 13118 1727204065.92670: done getting next task for host managed-node2 13118 1727204065.92672: ^ task is: TASK: Enable EPEL 7 13118 1727204065.92675: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204065.92678: getting variables 13118 1727204065.92679: in VariableManager get_vars() 13118 1727204065.92742: Calling all_inventory to load vars for managed-node2 13118 1727204065.92744: Calling groups_inventory to load vars for managed-node2 13118 1727204065.92747: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204065.92753: Calling all_plugins_play to load vars for managed-node2 13118 1727204065.92755: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204065.92757: Calling groups_plugins_play to load vars for managed-node2 13118 1727204065.92853: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204065.92967: done with get_vars() 13118 1727204065.92974: done getting variables 13118 1727204065.93012: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Enable EPEL 7] *********************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:32 Tuesday 24 September 2024 14:54:25 -0400 (0:00:00.015) 0:00:03.230 ***** 13118 1727204065.93040: entering _queue_task() for managed-node2/command 13118 1727204065.93262: worker is 1 (out of 1 available) 13118 1727204065.93277: exiting _queue_task() for managed-node2/command 13118 1727204065.93289: done queuing things up, now waiting for results queue to drain 13118 1727204065.93291: waiting for pending results... 13118 1727204065.93545: running TaskExecutor() for managed-node2/TASK: Enable EPEL 7 13118 1727204065.93666: in run() - task 0affcd87-79f5-56a3-0a64-0000000000ff 13118 1727204065.93687: variable 'ansible_search_path' from source: unknown 13118 1727204065.93695: variable 'ansible_search_path' from source: unknown 13118 1727204065.93740: calling self._execute() 13118 1727204065.93816: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.93831: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.93851: variable 'omit' from source: magic vars 13118 1727204065.94238: variable 'ansible_distribution' from source: facts 13118 1727204065.94256: Evaluated conditional (ansible_distribution in ['RedHat', 'CentOS']): True 13118 1727204065.94401: variable 'ansible_distribution_major_version' from source: facts 13118 1727204065.94414: Evaluated conditional (ansible_distribution_major_version in ['7', '8']): False 13118 1727204065.94422: when evaluation is False, skipping this task 13118 1727204065.94432: _execute() done 13118 1727204065.94440: dumping result to json 13118 1727204065.94449: done dumping result, returning 13118 1727204065.94459: done running TaskExecutor() for managed-node2/TASK: Enable EPEL 7 [0affcd87-79f5-56a3-0a64-0000000000ff] 13118 1727204065.94471: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000ff 13118 1727204065.94585: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000ff 13118 1727204065.94593: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_distribution_major_version in ['7', '8']", "skip_reason": "Conditional result was False" } 13118 1727204065.94653: no more pending results, returning what we have 13118 1727204065.94656: results queue empty 13118 1727204065.94657: checking for any_errors_fatal 13118 1727204065.94674: done checking for any_errors_fatal 13118 1727204065.94676: checking for max_fail_percentage 13118 1727204065.94678: done checking for max_fail_percentage 13118 1727204065.94679: checking to see if all hosts have failed and the running result is not ok 13118 1727204065.94679: done checking to see if all hosts have failed 13118 1727204065.94680: getting the remaining hosts for this loop 13118 1727204065.94681: done getting the remaining hosts for this loop 13118 1727204065.94686: getting the next task for host managed-node2 13118 1727204065.94694: done getting next task for host managed-node2 13118 1727204065.94696: ^ task is: TASK: Enable EPEL 8 13118 1727204065.94701: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204065.94705: getting variables 13118 1727204065.94707: in VariableManager get_vars() 13118 1727204065.94743: Calling all_inventory to load vars for managed-node2 13118 1727204065.94746: Calling groups_inventory to load vars for managed-node2 13118 1727204065.94750: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204065.94765: Calling all_plugins_play to load vars for managed-node2 13118 1727204065.94769: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204065.94773: Calling groups_plugins_play to load vars for managed-node2 13118 1727204065.94956: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204065.95176: done with get_vars() 13118 1727204065.95187: done getting variables 13118 1727204065.95268: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Enable EPEL 8] *********************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:37 Tuesday 24 September 2024 14:54:25 -0400 (0:00:00.022) 0:00:03.253 ***** 13118 1727204065.95303: entering _queue_task() for managed-node2/command 13118 1727204065.95779: worker is 1 (out of 1 available) 13118 1727204065.95791: exiting _queue_task() for managed-node2/command 13118 1727204065.95803: done queuing things up, now waiting for results queue to drain 13118 1727204065.95805: waiting for pending results... 13118 1727204065.96060: running TaskExecutor() for managed-node2/TASK: Enable EPEL 8 13118 1727204065.96180: in run() - task 0affcd87-79f5-56a3-0a64-000000000100 13118 1727204065.96199: variable 'ansible_search_path' from source: unknown 13118 1727204065.96206: variable 'ansible_search_path' from source: unknown 13118 1727204065.96251: calling self._execute() 13118 1727204065.96324: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.96442: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.96458: variable 'omit' from source: magic vars 13118 1727204065.96926: variable 'ansible_distribution' from source: facts 13118 1727204065.96947: Evaluated conditional (ansible_distribution in ['RedHat', 'CentOS']): True 13118 1727204065.97112: variable 'ansible_distribution_major_version' from source: facts 13118 1727204065.97125: Evaluated conditional (ansible_distribution_major_version in ['7', '8']): False 13118 1727204065.97140: when evaluation is False, skipping this task 13118 1727204065.97149: _execute() done 13118 1727204065.97157: dumping result to json 13118 1727204065.97166: done dumping result, returning 13118 1727204065.97177: done running TaskExecutor() for managed-node2/TASK: Enable EPEL 8 [0affcd87-79f5-56a3-0a64-000000000100] 13118 1727204065.97188: sending task result for task 0affcd87-79f5-56a3-0a64-000000000100 skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_distribution_major_version in ['7', '8']", "skip_reason": "Conditional result was False" } 13118 1727204065.97345: no more pending results, returning what we have 13118 1727204065.97349: results queue empty 13118 1727204065.97350: checking for any_errors_fatal 13118 1727204065.97356: done checking for any_errors_fatal 13118 1727204065.97357: checking for max_fail_percentage 13118 1727204065.97358: done checking for max_fail_percentage 13118 1727204065.97359: checking to see if all hosts have failed and the running result is not ok 13118 1727204065.97360: done checking to see if all hosts have failed 13118 1727204065.97361: getting the remaining hosts for this loop 13118 1727204065.97362: done getting the remaining hosts for this loop 13118 1727204065.97368: getting the next task for host managed-node2 13118 1727204065.97377: done getting next task for host managed-node2 13118 1727204065.97379: ^ task is: TASK: Enable EPEL 6 13118 1727204065.97383: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204065.97386: getting variables 13118 1727204065.97388: in VariableManager get_vars() 13118 1727204065.97468: Calling all_inventory to load vars for managed-node2 13118 1727204065.97471: Calling groups_inventory to load vars for managed-node2 13118 1727204065.97475: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204065.97488: Calling all_plugins_play to load vars for managed-node2 13118 1727204065.97491: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204065.97493: Calling groups_plugins_play to load vars for managed-node2 13118 1727204065.97652: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204065.98224: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000100 13118 1727204065.98230: WORKER PROCESS EXITING 13118 1727204065.98273: done with get_vars() 13118 1727204065.98283: done getting variables 13118 1727204065.98347: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Enable EPEL 6] *********************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:42 Tuesday 24 September 2024 14:54:25 -0400 (0:00:00.030) 0:00:03.283 ***** 13118 1727204065.98382: entering _queue_task() for managed-node2/copy 13118 1727204065.98650: worker is 1 (out of 1 available) 13118 1727204065.98666: exiting _queue_task() for managed-node2/copy 13118 1727204065.98679: done queuing things up, now waiting for results queue to drain 13118 1727204065.98681: waiting for pending results... 13118 1727204065.99547: running TaskExecutor() for managed-node2/TASK: Enable EPEL 6 13118 1727204065.99753: in run() - task 0affcd87-79f5-56a3-0a64-000000000102 13118 1727204065.99777: variable 'ansible_search_path' from source: unknown 13118 1727204065.99781: variable 'ansible_search_path' from source: unknown 13118 1727204065.99810: calling self._execute() 13118 1727204065.99900: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204065.99911: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204065.99925: variable 'omit' from source: magic vars 13118 1727204066.00334: variable 'ansible_distribution' from source: facts 13118 1727204066.00353: Evaluated conditional (ansible_distribution in ['RedHat', 'CentOS']): True 13118 1727204066.00486: variable 'ansible_distribution_major_version' from source: facts 13118 1727204066.00502: Evaluated conditional (ansible_distribution_major_version == '6'): False 13118 1727204066.00516: when evaluation is False, skipping this task 13118 1727204066.00526: _execute() done 13118 1727204066.00535: dumping result to json 13118 1727204066.00543: done dumping result, returning 13118 1727204066.00553: done running TaskExecutor() for managed-node2/TASK: Enable EPEL 6 [0affcd87-79f5-56a3-0a64-000000000102] 13118 1727204066.00567: sending task result for task 0affcd87-79f5-56a3-0a64-000000000102 skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_distribution_major_version == '6'", "skip_reason": "Conditional result was False" } 13118 1727204066.00731: no more pending results, returning what we have 13118 1727204066.00736: results queue empty 13118 1727204066.00736: checking for any_errors_fatal 13118 1727204066.00741: done checking for any_errors_fatal 13118 1727204066.00741: checking for max_fail_percentage 13118 1727204066.00743: done checking for max_fail_percentage 13118 1727204066.00744: checking to see if all hosts have failed and the running result is not ok 13118 1727204066.00745: done checking to see if all hosts have failed 13118 1727204066.00745: getting the remaining hosts for this loop 13118 1727204066.00747: done getting the remaining hosts for this loop 13118 1727204066.00750: getting the next task for host managed-node2 13118 1727204066.00760: done getting next task for host managed-node2 13118 1727204066.00762: ^ task is: TASK: Set network provider to 'nm' 13118 1727204066.00766: ^ state is: HOST STATE: block=2, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204066.00769: getting variables 13118 1727204066.00771: in VariableManager get_vars() 13118 1727204066.00800: Calling all_inventory to load vars for managed-node2 13118 1727204066.00803: Calling groups_inventory to load vars for managed-node2 13118 1727204066.00807: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204066.00821: Calling all_plugins_play to load vars for managed-node2 13118 1727204066.00823: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204066.00827: Calling groups_plugins_play to load vars for managed-node2 13118 1727204066.00989: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000102 13118 1727204066.00993: WORKER PROCESS EXITING 13118 1727204066.01013: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204066.01217: done with get_vars() 13118 1727204066.01234: done getting variables 13118 1727204066.01297: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Set network provider to 'nm'] ******************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tests_bond_nm.yml:13 Tuesday 24 September 2024 14:54:26 -0400 (0:00:00.029) 0:00:03.313 ***** 13118 1727204066.01326: entering _queue_task() for managed-node2/set_fact 13118 1727204066.01626: worker is 1 (out of 1 available) 13118 1727204066.01638: exiting _queue_task() for managed-node2/set_fact 13118 1727204066.01652: done queuing things up, now waiting for results queue to drain 13118 1727204066.01653: waiting for pending results... 13118 1727204066.01925: running TaskExecutor() for managed-node2/TASK: Set network provider to 'nm' 13118 1727204066.02018: in run() - task 0affcd87-79f5-56a3-0a64-000000000007 13118 1727204066.02037: variable 'ansible_search_path' from source: unknown 13118 1727204066.02079: calling self._execute() 13118 1727204066.03071: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204066.03085: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204066.03099: variable 'omit' from source: magic vars 13118 1727204066.03323: variable 'omit' from source: magic vars 13118 1727204066.03479: variable 'omit' from source: magic vars 13118 1727204066.03521: variable 'omit' from source: magic vars 13118 1727204066.03584: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204066.03640: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204066.03666: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204066.03693: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204066.03710: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204066.03747: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204066.03756: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204066.03765: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204066.03878: Set connection var ansible_timeout to 10 13118 1727204066.03906: Set connection var ansible_pipelining to False 13118 1727204066.03914: Set connection var ansible_connection to ssh 13118 1727204066.03923: Set connection var ansible_shell_executable to /bin/sh 13118 1727204066.03932: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204066.03938: Set connection var ansible_shell_type to sh 13118 1727204066.03966: variable 'ansible_shell_executable' from source: unknown 13118 1727204066.03975: variable 'ansible_connection' from source: unknown 13118 1727204066.03981: variable 'ansible_module_compression' from source: unknown 13118 1727204066.03989: variable 'ansible_shell_type' from source: unknown 13118 1727204066.04005: variable 'ansible_shell_executable' from source: unknown 13118 1727204066.04013: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204066.04021: variable 'ansible_pipelining' from source: unknown 13118 1727204066.04027: variable 'ansible_timeout' from source: unknown 13118 1727204066.04034: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204066.04189: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204066.04206: variable 'omit' from source: magic vars 13118 1727204066.04227: starting attempt loop 13118 1727204066.04236: running the handler 13118 1727204066.04252: handler run complete 13118 1727204066.04268: attempt loop complete, returning result 13118 1727204066.04275: _execute() done 13118 1727204066.04281: dumping result to json 13118 1727204066.04288: done dumping result, returning 13118 1727204066.04297: done running TaskExecutor() for managed-node2/TASK: Set network provider to 'nm' [0affcd87-79f5-56a3-0a64-000000000007] 13118 1727204066.04308: sending task result for task 0affcd87-79f5-56a3-0a64-000000000007 13118 1727204066.04422: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000007 ok: [managed-node2] => { "ansible_facts": { "network_provider": "nm" }, "changed": false } 13118 1727204066.04545: no more pending results, returning what we have 13118 1727204066.04550: results queue empty 13118 1727204066.04551: checking for any_errors_fatal 13118 1727204066.04556: done checking for any_errors_fatal 13118 1727204066.04557: checking for max_fail_percentage 13118 1727204066.04559: done checking for max_fail_percentage 13118 1727204066.04560: checking to see if all hosts have failed and the running result is not ok 13118 1727204066.04561: done checking to see if all hosts have failed 13118 1727204066.04561: getting the remaining hosts for this loop 13118 1727204066.04563: done getting the remaining hosts for this loop 13118 1727204066.04569: getting the next task for host managed-node2 13118 1727204066.04577: done getting next task for host managed-node2 13118 1727204066.04579: ^ task is: TASK: meta (flush_handlers) 13118 1727204066.04581: ^ state is: HOST STATE: block=3, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204066.04586: getting variables 13118 1727204066.04588: in VariableManager get_vars() 13118 1727204066.04614: Calling all_inventory to load vars for managed-node2 13118 1727204066.04617: Calling groups_inventory to load vars for managed-node2 13118 1727204066.04621: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204066.04632: Calling all_plugins_play to load vars for managed-node2 13118 1727204066.04636: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204066.04639: Calling groups_plugins_play to load vars for managed-node2 13118 1727204066.04840: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204066.05042: done with get_vars() 13118 1727204066.05052: done getting variables 13118 1727204066.05225: WORKER PROCESS EXITING 13118 1727204066.05251: in VariableManager get_vars() 13118 1727204066.05261: Calling all_inventory to load vars for managed-node2 13118 1727204066.05263: Calling groups_inventory to load vars for managed-node2 13118 1727204066.05268: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204066.05272: Calling all_plugins_play to load vars for managed-node2 13118 1727204066.05274: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204066.05277: Calling groups_plugins_play to load vars for managed-node2 13118 1727204066.05531: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204066.06148: done with get_vars() 13118 1727204066.06163: done queuing things up, now waiting for results queue to drain 13118 1727204066.06168: results queue empty 13118 1727204066.06168: checking for any_errors_fatal 13118 1727204066.06171: done checking for any_errors_fatal 13118 1727204066.06171: checking for max_fail_percentage 13118 1727204066.06172: done checking for max_fail_percentage 13118 1727204066.06173: checking to see if all hosts have failed and the running result is not ok 13118 1727204066.06174: done checking to see if all hosts have failed 13118 1727204066.06175: getting the remaining hosts for this loop 13118 1727204066.06175: done getting the remaining hosts for this loop 13118 1727204066.06178: getting the next task for host managed-node2 13118 1727204066.06182: done getting next task for host managed-node2 13118 1727204066.06184: ^ task is: TASK: meta (flush_handlers) 13118 1727204066.06185: ^ state is: HOST STATE: block=4, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204066.06194: getting variables 13118 1727204066.06196: in VariableManager get_vars() 13118 1727204066.06205: Calling all_inventory to load vars for managed-node2 13118 1727204066.06208: Calling groups_inventory to load vars for managed-node2 13118 1727204066.06210: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204066.06215: Calling all_plugins_play to load vars for managed-node2 13118 1727204066.06218: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204066.06221: Calling groups_plugins_play to load vars for managed-node2 13118 1727204066.07031: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204066.07581: done with get_vars() 13118 1727204066.07595: done getting variables 13118 1727204066.07768: in VariableManager get_vars() 13118 1727204066.07780: Calling all_inventory to load vars for managed-node2 13118 1727204066.07783: Calling groups_inventory to load vars for managed-node2 13118 1727204066.07785: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204066.07791: Calling all_plugins_play to load vars for managed-node2 13118 1727204066.07793: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204066.07796: Calling groups_plugins_play to load vars for managed-node2 13118 1727204066.08054: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204066.08810: done with get_vars() 13118 1727204066.08825: done queuing things up, now waiting for results queue to drain 13118 1727204066.08827: results queue empty 13118 1727204066.08828: checking for any_errors_fatal 13118 1727204066.08829: done checking for any_errors_fatal 13118 1727204066.08830: checking for max_fail_percentage 13118 1727204066.08832: done checking for max_fail_percentage 13118 1727204066.08832: checking to see if all hosts have failed and the running result is not ok 13118 1727204066.08833: done checking to see if all hosts have failed 13118 1727204066.08834: getting the remaining hosts for this loop 13118 1727204066.08835: done getting the remaining hosts for this loop 13118 1727204066.08838: getting the next task for host managed-node2 13118 1727204066.08956: done getting next task for host managed-node2 13118 1727204066.08957: ^ task is: None 13118 1727204066.08959: ^ state is: HOST STATE: block=5, task=0, rescue=0, always=0, handlers=0, run_state=5, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204066.08960: done queuing things up, now waiting for results queue to drain 13118 1727204066.08961: results queue empty 13118 1727204066.08962: checking for any_errors_fatal 13118 1727204066.08963: done checking for any_errors_fatal 13118 1727204066.08965: checking for max_fail_percentage 13118 1727204066.08967: done checking for max_fail_percentage 13118 1727204066.08967: checking to see if all hosts have failed and the running result is not ok 13118 1727204066.08968: done checking to see if all hosts have failed 13118 1727204066.08970: getting the next task for host managed-node2 13118 1727204066.08973: done getting next task for host managed-node2 13118 1727204066.08974: ^ task is: None 13118 1727204066.08975: ^ state is: HOST STATE: block=5, task=0, rescue=0, always=0, handlers=0, run_state=5, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204066.09029: in VariableManager get_vars() 13118 1727204066.09168: done with get_vars() 13118 1727204066.09178: in VariableManager get_vars() 13118 1727204066.09195: done with get_vars() 13118 1727204066.09199: variable 'omit' from source: magic vars 13118 1727204066.09232: in VariableManager get_vars() 13118 1727204066.09249: done with get_vars() 13118 1727204066.09389: variable 'omit' from source: magic vars PLAY [Play for testing bond connection] **************************************** 13118 1727204066.10892: Loading StrategyModule 'linear' from /usr/local/lib/python3.12/site-packages/ansible/plugins/strategy/linear.py (found_in_cache=True, class_only=False) 13118 1727204066.11145: getting the remaining hosts for this loop 13118 1727204066.11147: done getting the remaining hosts for this loop 13118 1727204066.11150: getting the next task for host managed-node2 13118 1727204066.11153: done getting next task for host managed-node2 13118 1727204066.11155: ^ task is: TASK: Gathering Facts 13118 1727204066.11157: ^ state is: HOST STATE: block=0, task=0, rescue=0, always=0, handlers=0, run_state=0, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=True, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204066.11159: getting variables 13118 1727204066.11160: in VariableManager get_vars() 13118 1727204066.11177: Calling all_inventory to load vars for managed-node2 13118 1727204066.11180: Calling groups_inventory to load vars for managed-node2 13118 1727204066.11182: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204066.11188: Calling all_plugins_play to load vars for managed-node2 13118 1727204066.11202: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204066.11207: Calling groups_plugins_play to load vars for managed-node2 13118 1727204066.11401: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204066.11807: done with get_vars() 13118 1727204066.11816: done getting variables 13118 1727204066.11956: Loading ActionModule 'gather_facts' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/gather_facts.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml:3 Tuesday 24 September 2024 14:54:26 -0400 (0:00:00.107) 0:00:03.420 ***** 13118 1727204066.12056: entering _queue_task() for managed-node2/gather_facts 13118 1727204066.12725: worker is 1 (out of 1 available) 13118 1727204066.12737: exiting _queue_task() for managed-node2/gather_facts 13118 1727204066.12749: done queuing things up, now waiting for results queue to drain 13118 1727204066.12751: waiting for pending results... 13118 1727204066.13373: running TaskExecutor() for managed-node2/TASK: Gathering Facts 13118 1727204066.13745: in run() - task 0affcd87-79f5-56a3-0a64-000000000128 13118 1727204066.13886: variable 'ansible_search_path' from source: unknown 13118 1727204066.13931: calling self._execute() 13118 1727204066.14015: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204066.14203: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204066.14381: variable 'omit' from source: magic vars 13118 1727204066.15192: variable 'ansible_distribution_major_version' from source: facts 13118 1727204066.15213: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204066.15224: variable 'omit' from source: magic vars 13118 1727204066.15257: variable 'omit' from source: magic vars 13118 1727204066.15435: variable 'omit' from source: magic vars 13118 1727204066.15554: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204066.15599: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204066.15694: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204066.15719: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204066.15740: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204066.15779: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204066.15882: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204066.15892: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204066.16000: Set connection var ansible_timeout to 10 13118 1727204066.16085: Set connection var ansible_pipelining to False 13118 1727204066.16192: Set connection var ansible_connection to ssh 13118 1727204066.16203: Set connection var ansible_shell_executable to /bin/sh 13118 1727204066.16212: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204066.16218: Set connection var ansible_shell_type to sh 13118 1727204066.16248: variable 'ansible_shell_executable' from source: unknown 13118 1727204066.16256: variable 'ansible_connection' from source: unknown 13118 1727204066.16262: variable 'ansible_module_compression' from source: unknown 13118 1727204066.16271: variable 'ansible_shell_type' from source: unknown 13118 1727204066.16278: variable 'ansible_shell_executable' from source: unknown 13118 1727204066.16284: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204066.16290: variable 'ansible_pipelining' from source: unknown 13118 1727204066.16297: variable 'ansible_timeout' from source: unknown 13118 1727204066.16378: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204066.17043: Loading ActionModule 'gather_facts' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/gather_facts.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204066.17060: variable 'omit' from source: magic vars 13118 1727204066.17073: starting attempt loop 13118 1727204066.17081: running the handler 13118 1727204066.17100: variable 'ansible_facts' from source: unknown 13118 1727204066.17123: _low_level_execute_command(): starting 13118 1727204066.17139: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204066.19097: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204066.19103: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204066.19124: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204066.19131: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204066.19304: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204066.19387: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204066.19394: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204066.19440: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204066.21007: stdout chunk (state=3): >>>/root <<< 13118 1727204066.21103: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204066.21184: stderr chunk (state=3): >>><<< 13118 1727204066.21188: stdout chunk (state=3): >>><<< 13118 1727204066.21276: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204066.21283: _low_level_execute_command(): starting 13118 1727204066.21287: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204066.2120962-13853-111278148815201 `" && echo ansible-tmp-1727204066.2120962-13853-111278148815201="` echo /root/.ansible/tmp/ansible-tmp-1727204066.2120962-13853-111278148815201 `" ) && sleep 0' 13118 1727204066.22835: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204066.22961: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204066.22968: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204066.23005: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204066.23009: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204066.23018: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204066.23021: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204066.23200: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204066.23204: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204066.23210: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204066.23266: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204066.25263: stdout chunk (state=3): >>>ansible-tmp-1727204066.2120962-13853-111278148815201=/root/.ansible/tmp/ansible-tmp-1727204066.2120962-13853-111278148815201 <<< 13118 1727204066.25358: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204066.25440: stderr chunk (state=3): >>><<< 13118 1727204066.25444: stdout chunk (state=3): >>><<< 13118 1727204066.25670: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204066.2120962-13853-111278148815201=/root/.ansible/tmp/ansible-tmp-1727204066.2120962-13853-111278148815201 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204066.25674: variable 'ansible_module_compression' from source: unknown 13118 1727204066.25677: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.setup-ZIP_DEFLATED 13118 1727204066.25679: variable 'ansible_facts' from source: unknown 13118 1727204066.25799: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204066.2120962-13853-111278148815201/AnsiballZ_setup.py 13118 1727204066.26354: Sending initial data 13118 1727204066.26367: Sent initial data (154 bytes) 13118 1727204066.30289: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204066.30293: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204066.30500: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204066.30505: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204066.30646: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204066.30976: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204066.30979: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204066.30982: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204066.32552: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204066.32631: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204066.32657: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmp8ihm9g_x /root/.ansible/tmp/ansible-tmp-1727204066.2120962-13853-111278148815201/AnsiballZ_setup.py <<< 13118 1727204066.32690: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204066.35661: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204066.35670: stderr chunk (state=3): >>><<< 13118 1727204066.35673: stdout chunk (state=3): >>><<< 13118 1727204066.35793: done transferring module to remote 13118 1727204066.35796: _low_level_execute_command(): starting 13118 1727204066.35799: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204066.2120962-13853-111278148815201/ /root/.ansible/tmp/ansible-tmp-1727204066.2120962-13853-111278148815201/AnsiballZ_setup.py && sleep 0' 13118 1727204066.37108: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204066.37114: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204066.37261: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204066.37267: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204066.37270: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204066.37499: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204066.37504: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204066.37568: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204066.39417: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204066.39421: stdout chunk (state=3): >>><<< 13118 1727204066.39424: stderr chunk (state=3): >>><<< 13118 1727204066.39542: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204066.39545: _low_level_execute_command(): starting 13118 1727204066.39547: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204066.2120962-13853-111278148815201/AnsiballZ_setup.py && sleep 0' 13118 1727204066.41056: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204066.41195: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204066.41213: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204066.41236: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204066.41285: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204066.41299: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204066.41313: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204066.41332: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204066.41343: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204066.41353: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204066.41365: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204066.41378: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204066.41413: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204066.41424: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204066.41437: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204066.41449: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204066.41633: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204066.41656: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204066.41675: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204066.41755: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204067.01086: stdout chunk (state=3): >>> {"ansible_facts": {"ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_lsb": {}, "ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "root", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAPleAC0mV69PNpLSbmzZvoLD9LsCBzX6IHRLXV1uktk0r66T6Y57EoVgflJTdo6yU0zTaJjonNzFmvC69tiRsCyywGjnvnBOvIH2vrgNGCUdVYPZbbtmQlJvol7NFFfyXQR4RSPqBKT67rYbCzbETM4j+bdDgTeDk6l7wXwz9RVvAAAAFQCuAyyjbOBDKyIW26LGcI9/nmWpHwAAAIEApIE1W6KQ7qs5kJXBdSaPoWaZUxuQhXkPWORFe7/MBn5SojDfxvJjFPo6t4QsovaCnm532Zghh1ZdB0pNm0vYcRbz3wMdfMucw/KHWt6ZEtI+sLwuMyhAVEXzmE34iXkyePtELiYzY6NyxuJ04IujI9UwD7ZnqFBHVFz529oXikIAAACBAPdUu+4Qo82CMcmrGD9vNUgtsts6GCjqBDuov8GJEALZ9ZNLlyVoNtBHLMQH9e0czLygyNGw/IDosRQkKdX4Vh4A7KXujTIOyytaN4JVJCuOBY/PeX4lreAO/UTTUJ27yT/J0Oy2Hbt+d8fZnTkZReRNPFCzvdb1nuPMG5nAyQtL", "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQCzkKXWiNuOrU77QQcZuT2T9XVh655Sh8Sv9vLWLa1uj7ceaNsB0TBiqvDFvYPENhdKceYaGAFU7sjqbmp5dlivYwPBiBWvcOgqnpBqrMG5SvP1RMiORpW6GupBLnUaMVjopPLIi0/CDlSl2eODcEnQI6BpxCCSedEKU9UrRrCFJy+6KPQXepPwKwPTd1TMzO8wpo57B5MYrjnquTNxMfgBkYsHB/V77d0tKq8qGBTkAPD8wEWLIcZOI+SyYEfCraQ95dOGAPRTFijnd7S15CugSlJ/vvcHSFXOlbgFzeNnU2jZneagkBfaOJch72opD3ebISSHCx1/kJvHN7MbksI+ljJa3Nw5LwP1XjUpT7dQMOZJDdVStXKp86K4XpWud+wMbQVVyU5QoFsCl7YTWWmSDRiPJOQI2myfizCT8i42rJ0WXm5OnqpHn1Jw4nGlcVnfgPQA/zxMldzReXdHnvriqKC9+97XgY6pj42YYP78PhOu1D2xH1AXmloNM+63VvU=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPT1h7wNcUomxtav688iXvnCnFqrHnEKf4gRaBY3w4BwbWOGxE8hq5snF9Tp+0agFeN/u980/y8BJWdWIO9Lz8I=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIPe8liWy3mh5GzC<<< 13118 1727204067.01109: stdout chunk (state=3): >>>z9W616J2ArXnLOjLOZSwfmBX3Q1SI", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Tuesday", "weekday_number": "2", "weeknumber": "39", "day": "24", "hour": "14", "minute": "54", "second": "26", "epoch": "1727204066", "epoch_int": "1727204066", "date": "2024-09-24", "time": "14:54:26", "iso8601_micro": "2024-09-24T18:54:26.726666Z", "iso8601": "2024-09-24T18:54:26Z", "iso8601_basic": "20240924T145426726666", "iso8601_basic_short": "20240924T145426", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_virtualization_type": "xen", "ansible_virtualization_role": "guest", "ansible_virtualization_tech_guest": ["xen"], "ansible_virtualization_tech_host": [], "ansible_local": {}, "ansible_iscsi_iqn": "", "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "9", "ansible_distribution_major_version": "9", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_hostnqn": "nqn.2014-08.org.nvmexpress:uuid:d5aef1ea-3141-48ae-bf33-0c6b351dd422", "ansible_python": {"version": {"major": 3, "minor": 9, "micro": 19, "releaselevel": "final", "serial": 0}, "version_info": [3, 9, 19, "final", 0], "executable": "/usr/bin/python3.9", "has_sslcontext": true, "type": "cpython"}, "ansible_fips": false, "ansible_system": "Linux", "ansible_kernel": "5.14.0-511.el9.x86_64", "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Thu Sep 19 06:52:39 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.9.19", "ansible_fqdn": "managed-node2", "ansible_hostname": "managed-node2", "ansible_nodename": "managed-node2", "ansible_domain": "", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "e28ab0e542474a869c23f7ace4640799", "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_apparmor": {"status": "disabled"}, "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-5.14.0-511.el9.x86_64", "root": "UUID=ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": "ttyS0,115200n8"}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-5.14.0-511.el9.x86_64", "root": "UUID=ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": ["tty0", "ttyS0,115200n8"]}, "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "ansible_env": {"SHELL": "/bin/bash", "PWD": "/root", "LOGNAME": "root", "XDG_SESSION_TYPE": "tty", "_": "/usr/bin/python3.9", "MOTD_SHOWN": "pam", "HOME": "/root", "LANG": "en_US.UTF-8", "LS_COLORS": "", "SSH_CONNECTION": "10.31.14.85 48676 10.31.13.78 22", "XDG_SESSION_CLASS": "user", "SELINUX_ROLE_REQUESTED": "", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "USER": "root", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "1", "XDG_SESSION_ID": "5", "XDG_RUNTIME_DIR": "/run/user/0", "SSH_CLIENT": "10.31.14.85 48676 22", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "which_declare": "declare -f", "PATH": "/root/.local/bin:/root/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "SELINUX_LEVEL_REQUESTED": "", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "SSH_TTY": "/dev/pts/0", "BASH_FUNC_which%%": "() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}"}, "ansible_loadavg": {"1m": 0.38, "5m": 0.26, "15m": 0.13}, "ansible_fibre_channel_wwn": [], "ansible_processor": ["0", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz", "1", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz"], "ansible_processor_count": 1, "ansible_processor_cores": 1, "ansible_processor_threads_per_core": 2, "ansible_processor_vcpus": 2, "ansible_processor_nproc": 2, "ansible_memtotal_mb": 3532, "ansible_memfree_mb": 2773, "ansible_swaptotal_mb": 0, "ansible_swapfree_mb": 0, "ansible_memory_mb": {"real": {"total": 3532, "used": 759, "free": 2773}, "nocache": {"free": 3241, "used": 291}, "swap": {"total": 0, "free": 0, "used": 0, "cached": 0}}, "ansible_bios_date": "08/24/2006", "ansible_bios_vendor": "Xen", "ansible_bios_version": "4.11.amazon", "ansible_board_asset_tag": "NA", "ansible_board_name": "NA", "ansible_board_serial": "NA", "ansible_board_vendor": "NA", "ansible_board_version": "NA", "ansible_chassis_asset_tag": "NA", "ansible_chassis_serial": "NA", "ansible_chassis_vendor": "Xen", "ansible_chassis_version": "NA", "ansible_form_factor": "Other", "ansible_product_name": "HVM domU", "ansible_product_serial": "ec243623-fa66-7445-44ba-1070930583a9", "ansible_product_uuid": "ec243623-fa66-7445-44ba-1070930583a9", "ansible_product_version": "4.11.amazon", "ansible_system_vendor": "Xen", "ansible_devices": {"xvda": {"virtual": 1, "links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "vendor": null, "model": null, "sas_address": null, "sas_device_handle": null, "removable": "0", "support_discard": "512", "partitions": {"xvda1": {"links": {"ids": [], "uuids": ["ad406aa3-aab4-4a6a-aa73-3e870a6316ae"], "labels": [], "masters": []}, "start": "2048", "sectors": "524285919", "sectorsize": 512, "size": "250.00 GB", "uuid": "ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "holders": []}}, "rotational": "0", "scheduler_mode": "mq-deadline", "sectors": "524288000", "sectorsize": "512", "size": "250.00 GB", "host": "", "holders": []}}, "ansible_device_links": {"ids": {}, "uuids": {"xvda1": ["ad406aa3-aab4-4a6a-aa73-3e870a6316ae"]}, "labels": {}, "masters": {}}, "ansible_uptime_seconds": 429, "ansible_lvm": "N/A", "ansible_mounts": [{"mount": "/", "device": "/dev/xvda1", "fstype": "xfs", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "dump": 0, "passno": 0, "size_total": 268367278080, "size_available": 264272691200, "block_size": 4096, "block_total": 65519355, "block_available": 64519700, "block_used": 999655, "inode_total": 131071472, "inode_available": 130998247, "inode_used": 73225, "uuid": "ad406aa3-aab4-4a6a-aa73-3e870a6316ae"}], "ansible_is_chroot": false, "ansible_interfaces": ["lo", "eth0"], "ansible_eth0": {"device": "eth0", "macaddress": "0a:ff:ff:f5:f2:b9", "mtu": 9001, "active": true, "module": "xen_netfront", "type": "ether", "pciid": "vif-0", "promisc": false, "ipv4": {"address": "10.31.13.78", "broadcast": "10.31.15.255", "netmask": "255.255.252.0", "network": "10.31.12.0", "prefix": "22"}, "ipv6": [{"address": "fe80::8ff:ffff:fef5:f2b9", "prefix": "64", "scope": "link"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "on [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv6": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on", "tx_scatter_gather_fraglist": "off [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "off [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "off [fixed]", "tx_lockless": "off [fixed]", "netns_local": "off [fixed]", "tx_gso_robust": "on [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_gso_list": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off", "loopback": "off [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_lo": {"device": "lo", "mtu": 65536, "active": true, "type": "loopback", "promisc": false, "ipv4": {"address": "127.0.0.1", "broadcast": "", "netmask": "255.0.0.0", "network": "127.0.0.0", "prefix": "8"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "on [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "on [fixed]", "tx_lockless": "on [fixed]", "netns_local": "on [fixed]", "tx_gso_robust": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "on", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "on", "tx_gso_list": "on", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off [fixed]", "loopback": "on [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_default_ipv4": {"gateway": "10.31.12.1", "interface": "eth0", "address": "10.31.13.78", "broadcast": "10.31.15.255", "netmask": "255.255.252.0", "network": "10.31.12.0", "prefix": "22", "macaddress": "0a:ff:ff:f5:f2:b9", "mtu": 9001, "type": "ether", "alias": "eth0"}, "ansible_default_ipv6": {}, "ansible_all_ipv4_addresses": ["10.31.13.78"], "ansible_all_ipv6_addresses": ["fe80::8ff:ffff:fef5:f2b9"], "ansible_locally_reachable_ips": {"ipv4": ["10.31.13.78", "127.0.0.0/8", "127.0.0.1"], "ipv6": ["::1", "fe80::8ff:ffff:fef5:f2b9"]}, "ansible_pkg_mgr": "dnf", "ansible_service_mgr": "systemd", "gather_subset": ["all"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["all"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} <<< 13118 1727204067.02788: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204067.02841: stderr chunk (state=3): >>><<< 13118 1727204067.02845: stdout chunk (state=3): >>><<< 13118 1727204067.03077: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_lsb": {}, "ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "root", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAPleAC0mV69PNpLSbmzZvoLD9LsCBzX6IHRLXV1uktk0r66T6Y57EoVgflJTdo6yU0zTaJjonNzFmvC69tiRsCyywGjnvnBOvIH2vrgNGCUdVYPZbbtmQlJvol7NFFfyXQR4RSPqBKT67rYbCzbETM4j+bdDgTeDk6l7wXwz9RVvAAAAFQCuAyyjbOBDKyIW26LGcI9/nmWpHwAAAIEApIE1W6KQ7qs5kJXBdSaPoWaZUxuQhXkPWORFe7/MBn5SojDfxvJjFPo6t4QsovaCnm532Zghh1ZdB0pNm0vYcRbz3wMdfMucw/KHWt6ZEtI+sLwuMyhAVEXzmE34iXkyePtELiYzY6NyxuJ04IujI9UwD7ZnqFBHVFz529oXikIAAACBAPdUu+4Qo82CMcmrGD9vNUgtsts6GCjqBDuov8GJEALZ9ZNLlyVoNtBHLMQH9e0czLygyNGw/IDosRQkKdX4Vh4A7KXujTIOyytaN4JVJCuOBY/PeX4lreAO/UTTUJ27yT/J0Oy2Hbt+d8fZnTkZReRNPFCzvdb1nuPMG5nAyQtL", "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQCzkKXWiNuOrU77QQcZuT2T9XVh655Sh8Sv9vLWLa1uj7ceaNsB0TBiqvDFvYPENhdKceYaGAFU7sjqbmp5dlivYwPBiBWvcOgqnpBqrMG5SvP1RMiORpW6GupBLnUaMVjopPLIi0/CDlSl2eODcEnQI6BpxCCSedEKU9UrRrCFJy+6KPQXepPwKwPTd1TMzO8wpo57B5MYrjnquTNxMfgBkYsHB/V77d0tKq8qGBTkAPD8wEWLIcZOI+SyYEfCraQ95dOGAPRTFijnd7S15CugSlJ/vvcHSFXOlbgFzeNnU2jZneagkBfaOJch72opD3ebISSHCx1/kJvHN7MbksI+ljJa3Nw5LwP1XjUpT7dQMOZJDdVStXKp86K4XpWud+wMbQVVyU5QoFsCl7YTWWmSDRiPJOQI2myfizCT8i42rJ0WXm5OnqpHn1Jw4nGlcVnfgPQA/zxMldzReXdHnvriqKC9+97XgY6pj42YYP78PhOu1D2xH1AXmloNM+63VvU=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPT1h7wNcUomxtav688iXvnCnFqrHnEKf4gRaBY3w4BwbWOGxE8hq5snF9Tp+0agFeN/u980/y8BJWdWIO9Lz8I=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIPe8liWy3mh5GzCz9W616J2ArXnLOjLOZSwfmBX3Q1SI", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Tuesday", "weekday_number": "2", "weeknumber": "39", "day": "24", "hour": "14", "minute": "54", "second": "26", "epoch": "1727204066", "epoch_int": "1727204066", "date": "2024-09-24", "time": "14:54:26", "iso8601_micro": "2024-09-24T18:54:26.726666Z", "iso8601": "2024-09-24T18:54:26Z", "iso8601_basic": "20240924T145426726666", "iso8601_basic_short": "20240924T145426", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_virtualization_type": "xen", "ansible_virtualization_role": "guest", "ansible_virtualization_tech_guest": ["xen"], "ansible_virtualization_tech_host": [], "ansible_local": {}, "ansible_iscsi_iqn": "", "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "9", "ansible_distribution_major_version": "9", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_hostnqn": "nqn.2014-08.org.nvmexpress:uuid:d5aef1ea-3141-48ae-bf33-0c6b351dd422", "ansible_python": {"version": {"major": 3, "minor": 9, "micro": 19, "releaselevel": "final", "serial": 0}, "version_info": [3, 9, 19, "final", 0], "executable": "/usr/bin/python3.9", "has_sslcontext": true, "type": "cpython"}, "ansible_fips": false, "ansible_system": "Linux", "ansible_kernel": "5.14.0-511.el9.x86_64", "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Thu Sep 19 06:52:39 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.9.19", "ansible_fqdn": "managed-node2", "ansible_hostname": "managed-node2", "ansible_nodename": "managed-node2", "ansible_domain": "", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "e28ab0e542474a869c23f7ace4640799", "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_apparmor": {"status": "disabled"}, "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-5.14.0-511.el9.x86_64", "root": "UUID=ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": "ttyS0,115200n8"}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-5.14.0-511.el9.x86_64", "root": "UUID=ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": ["tty0", "ttyS0,115200n8"]}, "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "ansible_env": {"SHELL": "/bin/bash", "PWD": "/root", "LOGNAME": "root", "XDG_SESSION_TYPE": "tty", "_": "/usr/bin/python3.9", "MOTD_SHOWN": "pam", "HOME": "/root", "LANG": "en_US.UTF-8", "LS_COLORS": "", "SSH_CONNECTION": "10.31.14.85 48676 10.31.13.78 22", "XDG_SESSION_CLASS": "user", "SELINUX_ROLE_REQUESTED": "", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "USER": "root", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "1", "XDG_SESSION_ID": "5", "XDG_RUNTIME_DIR": "/run/user/0", "SSH_CLIENT": "10.31.14.85 48676 22", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "which_declare": "declare -f", "PATH": "/root/.local/bin:/root/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "SELINUX_LEVEL_REQUESTED": "", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "SSH_TTY": "/dev/pts/0", "BASH_FUNC_which%%": "() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}"}, "ansible_loadavg": {"1m": 0.38, "5m": 0.26, "15m": 0.13}, "ansible_fibre_channel_wwn": [], "ansible_processor": ["0", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz", "1", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz"], "ansible_processor_count": 1, "ansible_processor_cores": 1, "ansible_processor_threads_per_core": 2, "ansible_processor_vcpus": 2, "ansible_processor_nproc": 2, "ansible_memtotal_mb": 3532, "ansible_memfree_mb": 2773, "ansible_swaptotal_mb": 0, "ansible_swapfree_mb": 0, "ansible_memory_mb": {"real": {"total": 3532, "used": 759, "free": 2773}, "nocache": {"free": 3241, "used": 291}, "swap": {"total": 0, "free": 0, "used": 0, "cached": 0}}, "ansible_bios_date": "08/24/2006", "ansible_bios_vendor": "Xen", "ansible_bios_version": "4.11.amazon", "ansible_board_asset_tag": "NA", "ansible_board_name": "NA", "ansible_board_serial": "NA", "ansible_board_vendor": "NA", "ansible_board_version": "NA", "ansible_chassis_asset_tag": "NA", "ansible_chassis_serial": "NA", "ansible_chassis_vendor": "Xen", "ansible_chassis_version": "NA", "ansible_form_factor": "Other", "ansible_product_name": "HVM domU", "ansible_product_serial": "ec243623-fa66-7445-44ba-1070930583a9", "ansible_product_uuid": "ec243623-fa66-7445-44ba-1070930583a9", "ansible_product_version": "4.11.amazon", "ansible_system_vendor": "Xen", "ansible_devices": {"xvda": {"virtual": 1, "links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "vendor": null, "model": null, "sas_address": null, "sas_device_handle": null, "removable": "0", "support_discard": "512", "partitions": {"xvda1": {"links": {"ids": [], "uuids": ["ad406aa3-aab4-4a6a-aa73-3e870a6316ae"], "labels": [], "masters": []}, "start": "2048", "sectors": "524285919", "sectorsize": 512, "size": "250.00 GB", "uuid": "ad406aa3-aab4-4a6a-aa73-3e870a6316ae", "holders": []}}, "rotational": "0", "scheduler_mode": "mq-deadline", "sectors": "524288000", "sectorsize": "512", "size": "250.00 GB", "host": "", "holders": []}}, "ansible_device_links": {"ids": {}, "uuids": {"xvda1": ["ad406aa3-aab4-4a6a-aa73-3e870a6316ae"]}, "labels": {}, "masters": {}}, "ansible_uptime_seconds": 429, "ansible_lvm": "N/A", "ansible_mounts": [{"mount": "/", "device": "/dev/xvda1", "fstype": "xfs", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "dump": 0, "passno": 0, "size_total": 268367278080, "size_available": 264272691200, "block_size": 4096, "block_total": 65519355, "block_available": 64519700, "block_used": 999655, "inode_total": 131071472, "inode_available": 130998247, "inode_used": 73225, "uuid": "ad406aa3-aab4-4a6a-aa73-3e870a6316ae"}], "ansible_is_chroot": false, "ansible_interfaces": ["lo", "eth0"], "ansible_eth0": {"device": "eth0", "macaddress": "0a:ff:ff:f5:f2:b9", "mtu": 9001, "active": true, "module": "xen_netfront", "type": "ether", "pciid": "vif-0", "promisc": false, "ipv4": {"address": "10.31.13.78", "broadcast": "10.31.15.255", "netmask": "255.255.252.0", "network": "10.31.12.0", "prefix": "22"}, "ipv6": [{"address": "fe80::8ff:ffff:fef5:f2b9", "prefix": "64", "scope": "link"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "on [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv6": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on", "tx_scatter_gather_fraglist": "off [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "off [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "off [fixed]", "tx_lockless": "off [fixed]", "netns_local": "off [fixed]", "tx_gso_robust": "on [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_gso_list": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off", "loopback": "off [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_lo": {"device": "lo", "mtu": 65536, "active": true, "type": "loopback", "promisc": false, "ipv4": {"address": "127.0.0.1", "broadcast": "", "netmask": "255.0.0.0", "network": "127.0.0.0", "prefix": "8"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "on [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "on [fixed]", "tx_lockless": "on [fixed]", "netns_local": "on [fixed]", "tx_gso_robust": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "on", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "on", "tx_gso_list": "on", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off [fixed]", "loopback": "on [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_default_ipv4": {"gateway": "10.31.12.1", "interface": "eth0", "address": "10.31.13.78", "broadcast": "10.31.15.255", "netmask": "255.255.252.0", "network": "10.31.12.0", "prefix": "22", "macaddress": "0a:ff:ff:f5:f2:b9", "mtu": 9001, "type": "ether", "alias": "eth0"}, "ansible_default_ipv6": {}, "ansible_all_ipv4_addresses": ["10.31.13.78"], "ansible_all_ipv6_addresses": ["fe80::8ff:ffff:fef5:f2b9"], "ansible_locally_reachable_ips": {"ipv4": ["10.31.13.78", "127.0.0.0/8", "127.0.0.1"], "ipv6": ["::1", "fe80::8ff:ffff:fef5:f2b9"]}, "ansible_pkg_mgr": "dnf", "ansible_service_mgr": "systemd", "gather_subset": ["all"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["all"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204067.03267: done with _execute_module (ansible.legacy.setup, {'_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.setup', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204066.2120962-13853-111278148815201/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204067.03301: _low_level_execute_command(): starting 13118 1727204067.03312: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204066.2120962-13853-111278148815201/ > /dev/null 2>&1 && sleep 0' 13118 1727204067.05161: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204067.05285: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204067.05307: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204067.05325: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204067.05374: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204067.05387: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204067.05407: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204067.05426: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204067.05443: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204067.05454: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204067.05469: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204067.05484: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204067.05500: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204067.05531: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204067.05544: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204067.05557: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204067.05750: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204067.05777: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204067.05793: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204067.05870: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204067.07775: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204067.07779: stdout chunk (state=3): >>><<< 13118 1727204067.07782: stderr chunk (state=3): >>><<< 13118 1727204067.07974: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204067.07977: handler run complete 13118 1727204067.07979: variable 'ansible_facts' from source: unknown 13118 1727204067.08056: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204067.08380: variable 'ansible_facts' from source: unknown 13118 1727204067.08681: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204067.09012: attempt loop complete, returning result 13118 1727204067.09071: _execute() done 13118 1727204067.09079: dumping result to json 13118 1727204067.09114: done dumping result, returning 13118 1727204067.09189: done running TaskExecutor() for managed-node2/TASK: Gathering Facts [0affcd87-79f5-56a3-0a64-000000000128] 13118 1727204067.09292: sending task result for task 0affcd87-79f5-56a3-0a64-000000000128 ok: [managed-node2] 13118 1727204067.10374: no more pending results, returning what we have 13118 1727204067.10377: results queue empty 13118 1727204067.10378: checking for any_errors_fatal 13118 1727204067.10380: done checking for any_errors_fatal 13118 1727204067.10381: checking for max_fail_percentage 13118 1727204067.10382: done checking for max_fail_percentage 13118 1727204067.10383: checking to see if all hosts have failed and the running result is not ok 13118 1727204067.10384: done checking to see if all hosts have failed 13118 1727204067.10385: getting the remaining hosts for this loop 13118 1727204067.10386: done getting the remaining hosts for this loop 13118 1727204067.10390: getting the next task for host managed-node2 13118 1727204067.10399: done getting next task for host managed-node2 13118 1727204067.10401: ^ task is: TASK: meta (flush_handlers) 13118 1727204067.10403: ^ state is: HOST STATE: block=1, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204067.10408: getting variables 13118 1727204067.10410: in VariableManager get_vars() 13118 1727204067.10451: Calling all_inventory to load vars for managed-node2 13118 1727204067.10454: Calling groups_inventory to load vars for managed-node2 13118 1727204067.10456: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204067.10470: Calling all_plugins_play to load vars for managed-node2 13118 1727204067.10473: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204067.10481: Calling groups_plugins_play to load vars for managed-node2 13118 1727204067.10647: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204067.11116: done with get_vars() 13118 1727204067.11325: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000128 13118 1727204067.11330: WORKER PROCESS EXITING 13118 1727204067.11335: done getting variables 13118 1727204067.11412: in VariableManager get_vars() 13118 1727204067.11432: Calling all_inventory to load vars for managed-node2 13118 1727204067.11435: Calling groups_inventory to load vars for managed-node2 13118 1727204067.11437: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204067.11442: Calling all_plugins_play to load vars for managed-node2 13118 1727204067.11444: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204067.11452: Calling groups_plugins_play to load vars for managed-node2 13118 1727204067.12354: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204067.13003: done with get_vars() 13118 1727204067.13017: done queuing things up, now waiting for results queue to drain 13118 1727204067.13019: results queue empty 13118 1727204067.13020: checking for any_errors_fatal 13118 1727204067.13024: done checking for any_errors_fatal 13118 1727204067.13025: checking for max_fail_percentage 13118 1727204067.13026: done checking for max_fail_percentage 13118 1727204067.13027: checking to see if all hosts have failed and the running result is not ok 13118 1727204067.13031: done checking to see if all hosts have failed 13118 1727204067.13031: getting the remaining hosts for this loop 13118 1727204067.13032: done getting the remaining hosts for this loop 13118 1727204067.13035: getting the next task for host managed-node2 13118 1727204067.13039: done getting next task for host managed-node2 13118 1727204067.13041: ^ task is: TASK: INIT Prepare setup 13118 1727204067.13043: ^ state is: HOST STATE: block=2, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204067.13045: getting variables 13118 1727204067.13046: in VariableManager get_vars() 13118 1727204067.13061: Calling all_inventory to load vars for managed-node2 13118 1727204067.13205: Calling groups_inventory to load vars for managed-node2 13118 1727204067.13210: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204067.13215: Calling all_plugins_play to load vars for managed-node2 13118 1727204067.13217: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204067.13220: Calling groups_plugins_play to load vars for managed-node2 13118 1727204067.13583: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204067.14331: done with get_vars() 13118 1727204067.14340: done getting variables 13118 1727204067.14539: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=False, class_only=True) TASK [INIT Prepare setup] ****************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml:15 Tuesday 24 September 2024 14:54:27 -0400 (0:00:01.025) 0:00:04.445 ***** 13118 1727204067.14566: entering _queue_task() for managed-node2/debug 13118 1727204067.14568: Creating lock for debug 13118 1727204067.15455: worker is 1 (out of 1 available) 13118 1727204067.15471: exiting _queue_task() for managed-node2/debug 13118 1727204067.15768: done queuing things up, now waiting for results queue to drain 13118 1727204067.15771: waiting for pending results... 13118 1727204067.16766: running TaskExecutor() for managed-node2/TASK: INIT Prepare setup 13118 1727204067.17258: in run() - task 0affcd87-79f5-56a3-0a64-00000000000b 13118 1727204067.17280: variable 'ansible_search_path' from source: unknown 13118 1727204067.17387: calling self._execute() 13118 1727204067.17634: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204067.17785: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204067.17804: variable 'omit' from source: magic vars 13118 1727204067.18943: variable 'ansible_distribution_major_version' from source: facts 13118 1727204067.19109: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204067.19212: variable 'omit' from source: magic vars 13118 1727204067.19242: variable 'omit' from source: magic vars 13118 1727204067.19335: variable 'omit' from source: magic vars 13118 1727204067.19533: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204067.19576: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204067.19666: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204067.19769: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204067.19787: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204067.19821: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204067.19832: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204067.19859: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204067.20085: Set connection var ansible_timeout to 10 13118 1727204067.20104: Set connection var ansible_pipelining to False 13118 1727204067.20177: Set connection var ansible_connection to ssh 13118 1727204067.20194: Set connection var ansible_shell_executable to /bin/sh 13118 1727204067.20204: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204067.20211: Set connection var ansible_shell_type to sh 13118 1727204067.20243: variable 'ansible_shell_executable' from source: unknown 13118 1727204067.20301: variable 'ansible_connection' from source: unknown 13118 1727204067.20312: variable 'ansible_module_compression' from source: unknown 13118 1727204067.20320: variable 'ansible_shell_type' from source: unknown 13118 1727204067.20327: variable 'ansible_shell_executable' from source: unknown 13118 1727204067.20337: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204067.20344: variable 'ansible_pipelining' from source: unknown 13118 1727204067.20351: variable 'ansible_timeout' from source: unknown 13118 1727204067.20358: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204067.20698: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204067.20715: variable 'omit' from source: magic vars 13118 1727204067.20739: starting attempt loop 13118 1727204067.20844: running the handler 13118 1727204067.20900: handler run complete 13118 1727204067.21013: attempt loop complete, returning result 13118 1727204067.21171: _execute() done 13118 1727204067.21182: dumping result to json 13118 1727204067.21191: done dumping result, returning 13118 1727204067.21202: done running TaskExecutor() for managed-node2/TASK: INIT Prepare setup [0affcd87-79f5-56a3-0a64-00000000000b] 13118 1727204067.21215: sending task result for task 0affcd87-79f5-56a3-0a64-00000000000b ok: [managed-node2] => {} MSG: ################################################## 13118 1727204067.21374: no more pending results, returning what we have 13118 1727204067.21380: results queue empty 13118 1727204067.21381: checking for any_errors_fatal 13118 1727204067.21382: done checking for any_errors_fatal 13118 1727204067.21383: checking for max_fail_percentage 13118 1727204067.21385: done checking for max_fail_percentage 13118 1727204067.21386: checking to see if all hosts have failed and the running result is not ok 13118 1727204067.21386: done checking to see if all hosts have failed 13118 1727204067.21387: getting the remaining hosts for this loop 13118 1727204067.21389: done getting the remaining hosts for this loop 13118 1727204067.21393: getting the next task for host managed-node2 13118 1727204067.21402: done getting next task for host managed-node2 13118 1727204067.21405: ^ task is: TASK: Install dnsmasq 13118 1727204067.21408: ^ state is: HOST STATE: block=2, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204067.21413: getting variables 13118 1727204067.21414: in VariableManager get_vars() 13118 1727204067.21461: Calling all_inventory to load vars for managed-node2 13118 1727204067.21466: Calling groups_inventory to load vars for managed-node2 13118 1727204067.21469: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204067.21480: Calling all_plugins_play to load vars for managed-node2 13118 1727204067.21483: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204067.21486: Calling groups_plugins_play to load vars for managed-node2 13118 1727204067.21712: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204067.21923: done with get_vars() 13118 1727204067.21937: done getting variables 13118 1727204067.21975: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000000b 13118 1727204067.21984: WORKER PROCESS EXITING 13118 1727204067.22143: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Install dnsmasq] ********************************************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:3 Tuesday 24 September 2024 14:54:27 -0400 (0:00:00.078) 0:00:04.523 ***** 13118 1727204067.22377: entering _queue_task() for managed-node2/package 13118 1727204067.22989: worker is 1 (out of 1 available) 13118 1727204067.23001: exiting _queue_task() for managed-node2/package 13118 1727204067.23013: done queuing things up, now waiting for results queue to drain 13118 1727204067.23014: waiting for pending results... 13118 1727204067.23866: running TaskExecutor() for managed-node2/TASK: Install dnsmasq 13118 1727204067.24162: in run() - task 0affcd87-79f5-56a3-0a64-00000000000f 13118 1727204067.24189: variable 'ansible_search_path' from source: unknown 13118 1727204067.24198: variable 'ansible_search_path' from source: unknown 13118 1727204067.24238: calling self._execute() 13118 1727204067.24421: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204067.24436: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204067.24483: variable 'omit' from source: magic vars 13118 1727204067.25297: variable 'ansible_distribution_major_version' from source: facts 13118 1727204067.25354: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204067.25369: variable 'omit' from source: magic vars 13118 1727204067.25490: variable 'omit' from source: magic vars 13118 1727204067.25924: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204067.31688: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204067.31898: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204067.32055: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204067.32170: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204067.32316: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204067.32516: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204067.32625: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204067.32785: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204067.32950: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204067.33044: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204067.33390: variable '__network_is_ostree' from source: set_fact 13118 1727204067.33480: variable 'omit' from source: magic vars 13118 1727204067.33596: variable 'omit' from source: magic vars 13118 1727204067.33647: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204067.33719: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204067.33817: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204067.33920: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204067.34021: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204067.34075: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204067.34134: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204067.34144: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204067.34422: Set connection var ansible_timeout to 10 13118 1727204067.34572: Set connection var ansible_pipelining to False 13118 1727204067.34605: Set connection var ansible_connection to ssh 13118 1727204067.34619: Set connection var ansible_shell_executable to /bin/sh 13118 1727204067.34680: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204067.34688: Set connection var ansible_shell_type to sh 13118 1727204067.34786: variable 'ansible_shell_executable' from source: unknown 13118 1727204067.34819: variable 'ansible_connection' from source: unknown 13118 1727204067.34833: variable 'ansible_module_compression' from source: unknown 13118 1727204067.34841: variable 'ansible_shell_type' from source: unknown 13118 1727204067.34848: variable 'ansible_shell_executable' from source: unknown 13118 1727204067.34873: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204067.34882: variable 'ansible_pipelining' from source: unknown 13118 1727204067.34894: variable 'ansible_timeout' from source: unknown 13118 1727204067.34930: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204067.35330: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204067.35374: variable 'omit' from source: magic vars 13118 1727204067.35384: starting attempt loop 13118 1727204067.35440: running the handler 13118 1727204067.35451: variable 'ansible_facts' from source: unknown 13118 1727204067.35458: variable 'ansible_facts' from source: unknown 13118 1727204067.35509: _low_level_execute_command(): starting 13118 1727204067.35583: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204067.38182: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204067.38188: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204067.38215: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204067.38219: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204067.38280: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204067.38386: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204067.38694: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204067.40348: stdout chunk (state=3): >>>/root <<< 13118 1727204067.40456: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204067.40546: stderr chunk (state=3): >>><<< 13118 1727204067.40549: stdout chunk (state=3): >>><<< 13118 1727204067.40674: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204067.40685: _low_level_execute_command(): starting 13118 1727204067.40687: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204067.4057193-13969-68754288407031 `" && echo ansible-tmp-1727204067.4057193-13969-68754288407031="` echo /root/.ansible/tmp/ansible-tmp-1727204067.4057193-13969-68754288407031 `" ) && sleep 0' 13118 1727204067.42751: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204067.42763: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204067.42771: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204067.42789: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204067.42944: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204067.43035: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204067.43045: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204067.43059: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204067.43069: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204067.43076: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204067.43085: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204067.43092: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204067.43103: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204067.43110: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204067.43116: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204067.43126: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204067.43202: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204067.43367: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204067.43378: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204067.43579: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204067.46031: stdout chunk (state=3): >>>ansible-tmp-1727204067.4057193-13969-68754288407031=/root/.ansible/tmp/ansible-tmp-1727204067.4057193-13969-68754288407031 <<< 13118 1727204067.46265: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204067.46269: stdout chunk (state=3): >>><<< 13118 1727204067.46278: stderr chunk (state=3): >>><<< 13118 1727204067.46301: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204067.4057193-13969-68754288407031=/root/.ansible/tmp/ansible-tmp-1727204067.4057193-13969-68754288407031 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204067.46334: variable 'ansible_module_compression' from source: unknown 13118 1727204067.46408: ANSIBALLZ: Using generic lock for ansible.legacy.dnf 13118 1727204067.46412: ANSIBALLZ: Acquiring lock 13118 1727204067.46415: ANSIBALLZ: Lock acquired: 140051944306976 13118 1727204067.46417: ANSIBALLZ: Creating module 13118 1727204067.90791: ANSIBALLZ: Writing module into payload 13118 1727204067.91085: ANSIBALLZ: Writing module 13118 1727204067.91121: ANSIBALLZ: Renaming module 13118 1727204067.91135: ANSIBALLZ: Done creating module 13118 1727204067.91157: variable 'ansible_facts' from source: unknown 13118 1727204067.91257: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204067.4057193-13969-68754288407031/AnsiballZ_dnf.py 13118 1727204067.91431: Sending initial data 13118 1727204067.91434: Sent initial data (151 bytes) 13118 1727204067.92836: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204067.92852: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204067.92870: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204067.92887: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204067.92937: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204067.92950: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204067.92968: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204067.92986: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204067.92997: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204067.93008: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204067.93027: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204067.93045: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204067.93061: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204067.93074: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204067.93084: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204067.93095: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204067.93178: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204067.93201: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204067.93220: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204067.93310: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 4 <<< 13118 1727204067.95854: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204067.95907: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204067.95947: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpf57ulkro /root/.ansible/tmp/ansible-tmp-1727204067.4057193-13969-68754288407031/AnsiballZ_dnf.py <<< 13118 1727204067.95962: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204067.97934: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204067.98060: stderr chunk (state=3): >>><<< 13118 1727204067.98071: stdout chunk (state=3): >>><<< 13118 1727204067.98075: done transferring module to remote 13118 1727204067.98077: _low_level_execute_command(): starting 13118 1727204067.98079: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204067.4057193-13969-68754288407031/ /root/.ansible/tmp/ansible-tmp-1727204067.4057193-13969-68754288407031/AnsiballZ_dnf.py && sleep 0' 13118 1727204068.00270: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204068.00397: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204068.00408: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204068.00424: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204068.00463: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204068.00472: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204068.00482: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204068.00503: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204068.00510: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204068.00517: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204068.00524: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204068.00534: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204068.00546: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204068.00553: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204068.00561: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204068.00574: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204068.00672: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204068.00943: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204068.00952: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204068.01154: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 4 <<< 13118 1727204068.03693: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204068.03697: stdout chunk (state=3): >>><<< 13118 1727204068.03704: stderr chunk (state=3): >>><<< 13118 1727204068.03735: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 4 debug2: Received exit status from master 0 13118 1727204068.03738: _low_level_execute_command(): starting 13118 1727204068.03741: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204067.4057193-13969-68754288407031/AnsiballZ_dnf.py && sleep 0' 13118 1727204068.05595: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204068.05686: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204068.05696: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204068.05709: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204068.05747: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204068.05754: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204068.05766: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204068.05782: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204068.05797: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204068.05802: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204068.05810: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204068.05820: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204068.05833: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204068.05840: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204068.05847: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204068.05909: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204068.05979: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204068.06111: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204068.06154: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204068.06975: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 4 <<< 13118 1727204070.75035: stdout chunk (state=3): >>> {"msg": "", "changed": true, "results": ["Installed: dnsmasq-2.85-16.el9.x86_64"], "rc": 0, "invocation": {"module_args": {"name": ["dnsmasq"], "state": "present", "allow_downgrade": false, "allowerasing": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "use_backend": "auto", "best": null, "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "nobest": null, "releasever": null}}} <<< 13118 1727204070.80288: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204070.80359: stderr chunk (state=3): >>><<< 13118 1727204070.80362: stdout chunk (state=3): >>><<< 13118 1727204070.80384: _low_level_execute_command() done: rc=0, stdout= {"msg": "", "changed": true, "results": ["Installed: dnsmasq-2.85-16.el9.x86_64"], "rc": 0, "invocation": {"module_args": {"name": ["dnsmasq"], "state": "present", "allow_downgrade": false, "allowerasing": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "use_backend": "auto", "best": null, "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "nobest": null, "releasever": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 4 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204070.80435: done with _execute_module (ansible.legacy.dnf, {'name': 'dnsmasq', 'state': 'present', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.dnf', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204067.4057193-13969-68754288407031/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204070.80439: _low_level_execute_command(): starting 13118 1727204070.80445: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204067.4057193-13969-68754288407031/ > /dev/null 2>&1 && sleep 0' 13118 1727204070.81061: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204070.81075: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204070.81085: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204070.81099: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204070.81142: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204070.81148: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204070.81158: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204070.81174: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204070.81181: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204070.81191: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204070.81195: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204070.81205: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204070.81217: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204070.81224: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204070.81233: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204070.81240: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204070.81313: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204070.81330: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204070.81338: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204070.81408: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204070.83369: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204070.83373: stderr chunk (state=3): >>><<< 13118 1727204070.83375: stdout chunk (state=3): >>><<< 13118 1727204070.83378: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204070.83380: handler run complete 13118 1727204070.83905: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204070.83909: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204070.83948: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204070.83986: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204070.84023: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204070.84097: variable '__install_status' from source: unknown 13118 1727204070.84120: Evaluated conditional (__install_status is success): True 13118 1727204070.84144: attempt loop complete, returning result 13118 1727204070.84151: _execute() done 13118 1727204070.84157: dumping result to json 13118 1727204070.84166: done dumping result, returning 13118 1727204070.84177: done running TaskExecutor() for managed-node2/TASK: Install dnsmasq [0affcd87-79f5-56a3-0a64-00000000000f] 13118 1727204070.84184: sending task result for task 0affcd87-79f5-56a3-0a64-00000000000f changed: [managed-node2] => { "attempts": 1, "changed": true, "rc": 0, "results": [ "Installed: dnsmasq-2.85-16.el9.x86_64" ] } 13118 1727204070.84794: no more pending results, returning what we have 13118 1727204070.84798: results queue empty 13118 1727204070.84799: checking for any_errors_fatal 13118 1727204070.84804: done checking for any_errors_fatal 13118 1727204070.84805: checking for max_fail_percentage 13118 1727204070.84808: done checking for max_fail_percentage 13118 1727204070.84808: checking to see if all hosts have failed and the running result is not ok 13118 1727204070.84809: done checking to see if all hosts have failed 13118 1727204070.84810: getting the remaining hosts for this loop 13118 1727204070.84812: done getting the remaining hosts for this loop 13118 1727204070.84816: getting the next task for host managed-node2 13118 1727204070.84823: done getting next task for host managed-node2 13118 1727204070.84825: ^ task is: TASK: Install pgrep, sysctl 13118 1727204070.84827: ^ state is: HOST STATE: block=2, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204070.84831: getting variables 13118 1727204070.84832: in VariableManager get_vars() 13118 1727204070.84873: Calling all_inventory to load vars for managed-node2 13118 1727204070.84876: Calling groups_inventory to load vars for managed-node2 13118 1727204070.84878: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204070.84890: Calling all_plugins_play to load vars for managed-node2 13118 1727204070.84893: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204070.84896: Calling groups_plugins_play to load vars for managed-node2 13118 1727204070.85072: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204070.85277: done with get_vars() 13118 1727204070.85288: done getting variables 13118 1727204070.85343: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Install pgrep, sysctl] *************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:17 Tuesday 24 September 2024 14:54:30 -0400 (0:00:03.630) 0:00:08.153 ***** 13118 1727204070.85381: entering _queue_task() for managed-node2/package 13118 1727204070.85400: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000000f 13118 1727204070.85410: WORKER PROCESS EXITING 13118 1727204070.86028: worker is 1 (out of 1 available) 13118 1727204070.86040: exiting _queue_task() for managed-node2/package 13118 1727204070.86052: done queuing things up, now waiting for results queue to drain 13118 1727204070.86053: waiting for pending results... 13118 1727204070.86921: running TaskExecutor() for managed-node2/TASK: Install pgrep, sysctl 13118 1727204070.87182: in run() - task 0affcd87-79f5-56a3-0a64-000000000010 13118 1727204070.87203: variable 'ansible_search_path' from source: unknown 13118 1727204070.87211: variable 'ansible_search_path' from source: unknown 13118 1727204070.87255: calling self._execute() 13118 1727204070.87337: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204070.87348: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204070.87360: variable 'omit' from source: magic vars 13118 1727204070.87720: variable 'ansible_distribution_major_version' from source: facts 13118 1727204070.87742: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204070.87867: variable 'ansible_os_family' from source: facts 13118 1727204070.87881: Evaluated conditional (ansible_os_family == 'RedHat'): True 13118 1727204070.88346: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204070.88611: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204070.88657: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204070.88694: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204070.88739: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204070.88816: variable 'ansible_distribution_major_version' from source: facts 13118 1727204070.88835: Evaluated conditional (ansible_distribution_major_version is version('6', '<=')): False 13118 1727204070.88842: when evaluation is False, skipping this task 13118 1727204070.88848: _execute() done 13118 1727204070.88854: dumping result to json 13118 1727204070.88860: done dumping result, returning 13118 1727204070.88871: done running TaskExecutor() for managed-node2/TASK: Install pgrep, sysctl [0affcd87-79f5-56a3-0a64-000000000010] 13118 1727204070.88880: sending task result for task 0affcd87-79f5-56a3-0a64-000000000010 skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_distribution_major_version is version('6', '<=')", "skip_reason": "Conditional result was False" } 13118 1727204070.89023: no more pending results, returning what we have 13118 1727204070.89026: results queue empty 13118 1727204070.89028: checking for any_errors_fatal 13118 1727204070.89034: done checking for any_errors_fatal 13118 1727204070.89035: checking for max_fail_percentage 13118 1727204070.89037: done checking for max_fail_percentage 13118 1727204070.89038: checking to see if all hosts have failed and the running result is not ok 13118 1727204070.89039: done checking to see if all hosts have failed 13118 1727204070.89040: getting the remaining hosts for this loop 13118 1727204070.89041: done getting the remaining hosts for this loop 13118 1727204070.89045: getting the next task for host managed-node2 13118 1727204070.89053: done getting next task for host managed-node2 13118 1727204070.89056: ^ task is: TASK: Install pgrep, sysctl 13118 1727204070.89058: ^ state is: HOST STATE: block=2, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204070.89062: getting variables 13118 1727204070.89065: in VariableManager get_vars() 13118 1727204070.89105: Calling all_inventory to load vars for managed-node2 13118 1727204070.89109: Calling groups_inventory to load vars for managed-node2 13118 1727204070.89111: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204070.89123: Calling all_plugins_play to load vars for managed-node2 13118 1727204070.89126: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204070.89130: Calling groups_plugins_play to load vars for managed-node2 13118 1727204070.89488: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204070.89952: done with get_vars() 13118 1727204070.89961: done getting variables 13118 1727204070.89992: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000010 13118 1727204070.89996: WORKER PROCESS EXITING 13118 1727204070.90029: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Install pgrep, sysctl] *************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:26 Tuesday 24 September 2024 14:54:30 -0400 (0:00:00.046) 0:00:08.200 ***** 13118 1727204070.90058: entering _queue_task() for managed-node2/package 13118 1727204070.90510: worker is 1 (out of 1 available) 13118 1727204070.90521: exiting _queue_task() for managed-node2/package 13118 1727204070.90532: done queuing things up, now waiting for results queue to drain 13118 1727204070.90533: waiting for pending results... 13118 1727204070.91227: running TaskExecutor() for managed-node2/TASK: Install pgrep, sysctl 13118 1727204070.91415: in run() - task 0affcd87-79f5-56a3-0a64-000000000011 13118 1727204070.91436: variable 'ansible_search_path' from source: unknown 13118 1727204070.91442: variable 'ansible_search_path' from source: unknown 13118 1727204070.91481: calling self._execute() 13118 1727204070.91589: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204070.91601: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204070.91623: variable 'omit' from source: magic vars 13118 1727204070.92018: variable 'ansible_distribution_major_version' from source: facts 13118 1727204070.92039: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204070.92163: variable 'ansible_os_family' from source: facts 13118 1727204070.92180: Evaluated conditional (ansible_os_family == 'RedHat'): True 13118 1727204070.92345: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204070.92634: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204070.92687: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204070.92737: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204070.92773: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204070.92861: variable 'ansible_distribution_major_version' from source: facts 13118 1727204070.92878: Evaluated conditional (ansible_distribution_major_version is version('7', '>=')): True 13118 1727204070.92887: variable 'omit' from source: magic vars 13118 1727204070.92944: variable 'omit' from source: magic vars 13118 1727204070.93103: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204070.95963: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204070.96039: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204070.96074: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204070.96106: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204070.96131: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204070.96216: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204070.96243: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204070.96271: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204070.96311: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204070.96325: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204070.96423: variable '__network_is_ostree' from source: set_fact 13118 1727204070.96431: variable 'omit' from source: magic vars 13118 1727204070.96458: variable 'omit' from source: magic vars 13118 1727204070.96485: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204070.96511: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204070.96527: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204070.96544: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204070.96554: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204070.96584: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204070.96588: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204070.96590: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204070.96687: Set connection var ansible_timeout to 10 13118 1727204070.96699: Set connection var ansible_pipelining to False 13118 1727204070.96702: Set connection var ansible_connection to ssh 13118 1727204070.96707: Set connection var ansible_shell_executable to /bin/sh 13118 1727204070.96712: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204070.96715: Set connection var ansible_shell_type to sh 13118 1727204070.96739: variable 'ansible_shell_executable' from source: unknown 13118 1727204070.96742: variable 'ansible_connection' from source: unknown 13118 1727204070.96745: variable 'ansible_module_compression' from source: unknown 13118 1727204070.96747: variable 'ansible_shell_type' from source: unknown 13118 1727204070.96749: variable 'ansible_shell_executable' from source: unknown 13118 1727204070.96752: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204070.96754: variable 'ansible_pipelining' from source: unknown 13118 1727204070.96758: variable 'ansible_timeout' from source: unknown 13118 1727204070.96763: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204070.96853: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204070.96863: variable 'omit' from source: magic vars 13118 1727204070.96871: starting attempt loop 13118 1727204070.96874: running the handler 13118 1727204070.96880: variable 'ansible_facts' from source: unknown 13118 1727204070.96883: variable 'ansible_facts' from source: unknown 13118 1727204070.96916: _low_level_execute_command(): starting 13118 1727204070.96923: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204070.97626: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204070.97639: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204070.97650: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204070.97665: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204070.97708: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204070.97714: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204070.97723: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204070.97737: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204070.97750: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204070.97753: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204070.97755: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204070.97766: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204070.97780: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204070.97787: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204070.97794: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204070.97803: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204070.97876: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204070.97892: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204070.97895: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204070.97967: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204070.99583: stdout chunk (state=3): >>>/root <<< 13118 1727204070.99771: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204070.99775: stdout chunk (state=3): >>><<< 13118 1727204070.99793: stderr chunk (state=3): >>><<< 13118 1727204070.99908: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204070.99912: _low_level_execute_command(): starting 13118 1727204070.99915: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204070.9981391-14180-113127497021375 `" && echo ansible-tmp-1727204070.9981391-14180-113127497021375="` echo /root/.ansible/tmp/ansible-tmp-1727204070.9981391-14180-113127497021375 `" ) && sleep 0' 13118 1727204071.00978: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204071.00982: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204071.01027: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204071.01033: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204071.01035: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204071.01038: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204071.01112: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204071.01115: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204071.01130: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204071.01205: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204071.03138: stdout chunk (state=3): >>>ansible-tmp-1727204070.9981391-14180-113127497021375=/root/.ansible/tmp/ansible-tmp-1727204070.9981391-14180-113127497021375 <<< 13118 1727204071.03382: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204071.03386: stdout chunk (state=3): >>><<< 13118 1727204071.03388: stderr chunk (state=3): >>><<< 13118 1727204071.03400: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204070.9981391-14180-113127497021375=/root/.ansible/tmp/ansible-tmp-1727204070.9981391-14180-113127497021375 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204071.03437: variable 'ansible_module_compression' from source: unknown 13118 1727204071.03501: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.dnf-ZIP_DEFLATED 13118 1727204071.03546: variable 'ansible_facts' from source: unknown 13118 1727204071.03646: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204070.9981391-14180-113127497021375/AnsiballZ_dnf.py 13118 1727204071.03844: Sending initial data 13118 1727204071.03848: Sent initial data (152 bytes) 13118 1727204071.05229: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204071.05238: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204071.05281: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204071.05285: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration <<< 13118 1727204071.05298: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204071.05304: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204071.05313: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204071.05330: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204071.05412: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204071.05435: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204071.05501: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204071.07284: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204071.07329: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204071.07361: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpkk3_addg /root/.ansible/tmp/ansible-tmp-1727204070.9981391-14180-113127497021375/AnsiballZ_dnf.py <<< 13118 1727204071.07378: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204071.09074: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204071.09258: stderr chunk (state=3): >>><<< 13118 1727204071.09261: stdout chunk (state=3): >>><<< 13118 1727204071.09269: done transferring module to remote 13118 1727204071.09272: _low_level_execute_command(): starting 13118 1727204071.09275: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204070.9981391-14180-113127497021375/ /root/.ansible/tmp/ansible-tmp-1727204070.9981391-14180-113127497021375/AnsiballZ_dnf.py && sleep 0' 13118 1727204071.09919: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204071.09923: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204071.09961: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204071.09966: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204071.09969: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204071.09971: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204071.10053: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204071.10057: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204071.10108: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204071.12021: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204071.12025: stdout chunk (state=3): >>><<< 13118 1727204071.12028: stderr chunk (state=3): >>><<< 13118 1727204071.12135: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204071.12139: _low_level_execute_command(): starting 13118 1727204071.12142: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204070.9981391-14180-113127497021375/AnsiballZ_dnf.py && sleep 0' 13118 1727204071.12758: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204071.12777: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204071.12793: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204071.12812: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204071.12857: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204071.12872: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204071.12888: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204071.12906: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204071.12919: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204071.12930: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204071.12943: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204071.12957: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204071.12995: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204071.13009: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204071.13020: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204071.13034: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204071.13110: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204071.13134: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204071.13152: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204071.13230: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204072.10251: stdout chunk (state=3): >>> {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["procps-ng"], "state": "present", "allow_downgrade": false, "allowerasing": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "use_backend": "auto", "best": null, "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "nobest": null, "releasever": null}}} <<< 13118 1727204072.14337: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204072.14377: stderr chunk (state=3): >>><<< 13118 1727204072.14381: stdout chunk (state=3): >>><<< 13118 1727204072.14398: _low_level_execute_command() done: rc=0, stdout= {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["procps-ng"], "state": "present", "allow_downgrade": false, "allowerasing": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "use_backend": "auto", "best": null, "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "nobest": null, "releasever": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204072.14434: done with _execute_module (ansible.legacy.dnf, {'name': 'procps-ng', 'state': 'present', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.dnf', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204070.9981391-14180-113127497021375/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204072.14441: _low_level_execute_command(): starting 13118 1727204072.14446: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204070.9981391-14180-113127497021375/ > /dev/null 2>&1 && sleep 0' 13118 1727204072.14950: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204072.14980: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204072.15005: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204072.15024: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204072.15078: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204072.15082: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204072.15084: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204072.15087: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204072.15157: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204072.15161: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204072.15163: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204072.15224: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204072.17046: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204072.17153: stderr chunk (state=3): >>><<< 13118 1727204072.17166: stdout chunk (state=3): >>><<< 13118 1727204072.17378: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204072.17382: handler run complete 13118 1727204072.17384: attempt loop complete, returning result 13118 1727204072.17386: _execute() done 13118 1727204072.17388: dumping result to json 13118 1727204072.17390: done dumping result, returning 13118 1727204072.17391: done running TaskExecutor() for managed-node2/TASK: Install pgrep, sysctl [0affcd87-79f5-56a3-0a64-000000000011] 13118 1727204072.17393: sending task result for task 0affcd87-79f5-56a3-0a64-000000000011 13118 1727204072.17460: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000011 13118 1727204072.17466: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do 13118 1727204072.17549: no more pending results, returning what we have 13118 1727204072.17552: results queue empty 13118 1727204072.17553: checking for any_errors_fatal 13118 1727204072.17561: done checking for any_errors_fatal 13118 1727204072.17561: checking for max_fail_percentage 13118 1727204072.17565: done checking for max_fail_percentage 13118 1727204072.17566: checking to see if all hosts have failed and the running result is not ok 13118 1727204072.17566: done checking to see if all hosts have failed 13118 1727204072.17567: getting the remaining hosts for this loop 13118 1727204072.17569: done getting the remaining hosts for this loop 13118 1727204072.17573: getting the next task for host managed-node2 13118 1727204072.17580: done getting next task for host managed-node2 13118 1727204072.17582: ^ task is: TASK: Create test interfaces 13118 1727204072.17585: ^ state is: HOST STATE: block=2, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204072.17593: getting variables 13118 1727204072.17595: in VariableManager get_vars() 13118 1727204072.17636: Calling all_inventory to load vars for managed-node2 13118 1727204072.17639: Calling groups_inventory to load vars for managed-node2 13118 1727204072.17641: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204072.17653: Calling all_plugins_play to load vars for managed-node2 13118 1727204072.17655: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204072.17658: Calling groups_plugins_play to load vars for managed-node2 13118 1727204072.17951: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204072.18348: done with get_vars() 13118 1727204072.18361: done getting variables 13118 1727204072.18472: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=False, class_only=True) TASK [Create test interfaces] ************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:35 Tuesday 24 September 2024 14:54:32 -0400 (0:00:01.284) 0:00:09.485 ***** 13118 1727204072.18503: entering _queue_task() for managed-node2/shell 13118 1727204072.18505: Creating lock for shell 13118 1727204072.18801: worker is 1 (out of 1 available) 13118 1727204072.18814: exiting _queue_task() for managed-node2/shell 13118 1727204072.18827: done queuing things up, now waiting for results queue to drain 13118 1727204072.18828: waiting for pending results... 13118 1727204072.19125: running TaskExecutor() for managed-node2/TASK: Create test interfaces 13118 1727204072.19255: in run() - task 0affcd87-79f5-56a3-0a64-000000000012 13118 1727204072.19279: variable 'ansible_search_path' from source: unknown 13118 1727204072.19288: variable 'ansible_search_path' from source: unknown 13118 1727204072.19337: calling self._execute() 13118 1727204072.19440: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204072.19452: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204072.19467: variable 'omit' from source: magic vars 13118 1727204072.19977: variable 'ansible_distribution_major_version' from source: facts 13118 1727204072.19995: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204072.20008: variable 'omit' from source: magic vars 13118 1727204072.20072: variable 'omit' from source: magic vars 13118 1727204072.20742: variable 'dhcp_interface1' from source: play vars 13118 1727204072.20755: variable 'dhcp_interface2' from source: play vars 13118 1727204072.20848: variable 'omit' from source: magic vars 13118 1727204072.20959: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204072.21077: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204072.21159: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204072.21185: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204072.21252: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204072.21299: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204072.21351: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204072.21364: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204072.21604: Set connection var ansible_timeout to 10 13118 1727204072.21692: Set connection var ansible_pipelining to False 13118 1727204072.21707: Set connection var ansible_connection to ssh 13118 1727204072.21718: Set connection var ansible_shell_executable to /bin/sh 13118 1727204072.21728: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204072.21735: Set connection var ansible_shell_type to sh 13118 1727204072.21767: variable 'ansible_shell_executable' from source: unknown 13118 1727204072.21798: variable 'ansible_connection' from source: unknown 13118 1727204072.21827: variable 'ansible_module_compression' from source: unknown 13118 1727204072.21898: variable 'ansible_shell_type' from source: unknown 13118 1727204072.21910: variable 'ansible_shell_executable' from source: unknown 13118 1727204072.21923: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204072.21936: variable 'ansible_pipelining' from source: unknown 13118 1727204072.21943: variable 'ansible_timeout' from source: unknown 13118 1727204072.21951: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204072.22297: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204072.22313: variable 'omit' from source: magic vars 13118 1727204072.22328: starting attempt loop 13118 1727204072.22339: running the handler 13118 1727204072.22359: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204072.22388: _low_level_execute_command(): starting 13118 1727204072.22400: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204072.24313: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204072.24338: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204072.24354: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204072.24378: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204072.24430: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204072.24449: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204072.24474: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204072.24495: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204072.24507: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204072.24517: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204072.24532: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204072.24557: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204072.24575: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204072.24588: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204072.24598: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204072.24611: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204072.24810: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204072.24834: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204072.24852: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204072.24943: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204072.26619: stdout chunk (state=3): >>>/root <<< 13118 1727204072.26819: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204072.26823: stdout chunk (state=3): >>><<< 13118 1727204072.26825: stderr chunk (state=3): >>><<< 13118 1727204072.26955: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204072.26969: _low_level_execute_command(): starting 13118 1727204072.26972: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204072.2685003-14388-117237243238353 `" && echo ansible-tmp-1727204072.2685003-14388-117237243238353="` echo /root/.ansible/tmp/ansible-tmp-1727204072.2685003-14388-117237243238353 `" ) && sleep 0' 13118 1727204072.28255: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204072.28262: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204072.28293: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204072.28297: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204072.28359: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204072.28839: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204072.28874: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204072.30680: stdout chunk (state=3): >>>ansible-tmp-1727204072.2685003-14388-117237243238353=/root/.ansible/tmp/ansible-tmp-1727204072.2685003-14388-117237243238353 <<< 13118 1727204072.30794: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204072.30881: stderr chunk (state=3): >>><<< 13118 1727204072.30885: stdout chunk (state=3): >>><<< 13118 1727204072.31078: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204072.2685003-14388-117237243238353=/root/.ansible/tmp/ansible-tmp-1727204072.2685003-14388-117237243238353 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204072.31082: variable 'ansible_module_compression' from source: unknown 13118 1727204072.31085: ANSIBALLZ: Using generic lock for ansible.legacy.command 13118 1727204072.31087: ANSIBALLZ: Acquiring lock 13118 1727204072.31089: ANSIBALLZ: Lock acquired: 140051944306976 13118 1727204072.31092: ANSIBALLZ: Creating module 13118 1727204072.54237: ANSIBALLZ: Writing module into payload 13118 1727204072.54781: ANSIBALLZ: Writing module 13118 1727204072.54836: ANSIBALLZ: Renaming module 13118 1727204072.54876: ANSIBALLZ: Done creating module 13118 1727204072.54881: variable 'ansible_facts' from source: unknown 13118 1727204072.54969: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204072.2685003-14388-117237243238353/AnsiballZ_command.py 13118 1727204072.55116: Sending initial data 13118 1727204072.55120: Sent initial data (156 bytes) 13118 1727204072.56086: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204072.56095: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204072.56105: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204072.56123: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204072.56160: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204072.56167: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204072.56178: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204072.56191: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204072.56198: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204072.56205: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204072.56212: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204072.56222: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204072.56236: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204072.56243: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204072.56249: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204072.56259: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204072.56334: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204072.56350: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204072.56360: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204072.56566: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204072.58345: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204072.58387: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204072.58428: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpgnfv7ypp /root/.ansible/tmp/ansible-tmp-1727204072.2685003-14388-117237243238353/AnsiballZ_command.py <<< 13118 1727204072.58465: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204072.59580: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204072.59746: stderr chunk (state=3): >>><<< 13118 1727204072.59750: stdout chunk (state=3): >>><<< 13118 1727204072.59773: done transferring module to remote 13118 1727204072.59786: _low_level_execute_command(): starting 13118 1727204072.59791: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204072.2685003-14388-117237243238353/ /root/.ansible/tmp/ansible-tmp-1727204072.2685003-14388-117237243238353/AnsiballZ_command.py && sleep 0' 13118 1727204072.60909: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204072.60914: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204072.60960: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204072.60965: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204072.60982: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204072.60988: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204072.61058: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204072.61073: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204072.61081: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204072.61144: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204072.62956: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204072.62961: stderr chunk (state=3): >>><<< 13118 1727204072.62966: stdout chunk (state=3): >>><<< 13118 1727204072.62983: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204072.62986: _low_level_execute_command(): starting 13118 1727204072.62990: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204072.2685003-14388-117237243238353/AnsiballZ_command.py && sleep 0' 13118 1727204072.63867: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204072.63872: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204072.63914: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204072.63918: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration <<< 13118 1727204072.63934: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204072.63938: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204072.63957: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204072.64031: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204072.64045: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204072.64054: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204072.64137: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.00032: stdout chunk (state=3): >>> <<< 13118 1727204074.00066: stdout chunk (state=3): >>>{"changed": true, "stdout": "", "stderr": "+ exec\n+ ip link add test1 type veth peer name test1p\n+ ip link add test2 type veth peer name test2p\n++ pgrep NetworkManager\n+ '[' -n 6823 ']'\n+ nmcli d set test1 managed true\n+ nmcli d set test2 managed true\n+ nmcli d set test1p managed false\n+ nmcli d set test2p managed false\n+ ip link set test1p up\n+ ip link set test2p up\n+ ip link add name testbr type bridge forward_delay 0\n++ pgrep NetworkManager\n+ '[' -n 6823 ']'\n+ nmcli d set testbr managed false\n+ ip link set testbr up\n+ timer=0\n+ ip addr show testbr\n+ grep -q 'inet [1-9]'\n+ let timer+=1\n+ '[' 1 -eq 30 ']'\n+ sleep 1\n+ rc=0\n+ ip addr add 192.0.2.1/24 dev testbr\n+ '[' 0 '!=' 0 ']'\n+ ip -6 addr add 2001:DB8::1/32 dev testbr\n+ '[' 0 '!=' 0 ']'\n+ grep -q 'inet [1-9]'\n+ ip addr show testbr\n+ grep 'release 6' /etc/redhat-release\n+ ip link set test1p master testbr\n+ ip link set test2p master testbr\n+ systemctl is-active firewalld\ninactive\n+ dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.co<<< 13118 1727204074.00073: stdout chunk (state=3): >>>m/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "start": "2024-09-24 14:54:32.771383", "end": "2024-09-24 14:54:33.999246", "delta": "0:00:01.227863", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}}<<< 13118 1727204074.00084: stdout chunk (state=3): >>> <<< 13118 1727204074.01461: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204074.01526: stderr chunk (state=3): >>><<< 13118 1727204074.01531: stdout chunk (state=3): >>><<< 13118 1727204074.01553: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "+ exec\n+ ip link add test1 type veth peer name test1p\n+ ip link add test2 type veth peer name test2p\n++ pgrep NetworkManager\n+ '[' -n 6823 ']'\n+ nmcli d set test1 managed true\n+ nmcli d set test2 managed true\n+ nmcli d set test1p managed false\n+ nmcli d set test2p managed false\n+ ip link set test1p up\n+ ip link set test2p up\n+ ip link add name testbr type bridge forward_delay 0\n++ pgrep NetworkManager\n+ '[' -n 6823 ']'\n+ nmcli d set testbr managed false\n+ ip link set testbr up\n+ timer=0\n+ ip addr show testbr\n+ grep -q 'inet [1-9]'\n+ let timer+=1\n+ '[' 1 -eq 30 ']'\n+ sleep 1\n+ rc=0\n+ ip addr add 192.0.2.1/24 dev testbr\n+ '[' 0 '!=' 0 ']'\n+ ip -6 addr add 2001:DB8::1/32 dev testbr\n+ '[' 0 '!=' 0 ']'\n+ grep -q 'inet [1-9]'\n+ ip addr show testbr\n+ grep 'release 6' /etc/redhat-release\n+ ip link set test1p master testbr\n+ ip link set test2p master testbr\n+ systemctl is-active firewalld\ninactive\n+ dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "start": "2024-09-24 14:54:32.771383", "end": "2024-09-24 14:54:33.999246", "delta": "0:00:01.227863", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204074.01597: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n "$(pgrep NetworkManager)" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the \'testbr\' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n "$(pgrep NetworkManager)" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q \'inet [1-9]\'\ndo\n let "timer+=1"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc="$?"\n if [ "$rc" != 0 ]; then\n echo NOTICE - could not add testbr - error code "$rc"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc="$?"\n if [ "$rc" != 0 ]; then\n echo NOTICE - could not add testbr - error code "$rc"\n continue\n fi\ndone\n\nif grep \'release 6\' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo \'interface testbr {\' > /etc/radvd.conf\n echo \' AdvSendAdvert on;\' >> /etc/radvd.conf\n echo \' prefix 2001:DB8::/64 { \' >> /etc/radvd.conf\n echo \' AdvOnLink on; }; \' >> /etc/radvd.conf\n echo \' }; \' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service="$service"; then\n firewall-cmd --add-service "$service"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204072.2685003-14388-117237243238353/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204074.01604: _low_level_execute_command(): starting 13118 1727204074.01609: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204072.2685003-14388-117237243238353/ > /dev/null 2>&1 && sleep 0' 13118 1727204074.02084: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.02090: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.02123: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.02138: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204074.02149: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.02200: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.02211: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.02258: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.04059: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204074.04117: stderr chunk (state=3): >>><<< 13118 1727204074.04120: stdout chunk (state=3): >>><<< 13118 1727204074.04135: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204074.04141: handler run complete 13118 1727204074.04158: Evaluated conditional (False): False 13118 1727204074.04170: attempt loop complete, returning result 13118 1727204074.04173: _execute() done 13118 1727204074.04176: dumping result to json 13118 1727204074.04182: done dumping result, returning 13118 1727204074.04188: done running TaskExecutor() for managed-node2/TASK: Create test interfaces [0affcd87-79f5-56a3-0a64-000000000012] 13118 1727204074.04193: sending task result for task 0affcd87-79f5-56a3-0a64-000000000012 13118 1727204074.04298: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000012 13118 1727204074.04301: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "cmd": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "delta": "0:00:01.227863", "end": "2024-09-24 14:54:33.999246", "rc": 0, "start": "2024-09-24 14:54:32.771383" } STDERR: + exec + ip link add test1 type veth peer name test1p + ip link add test2 type veth peer name test2p ++ pgrep NetworkManager + '[' -n 6823 ']' + nmcli d set test1 managed true + nmcli d set test2 managed true + nmcli d set test1p managed false + nmcli d set test2p managed false + ip link set test1p up + ip link set test2p up + ip link add name testbr type bridge forward_delay 0 ++ pgrep NetworkManager + '[' -n 6823 ']' + nmcli d set testbr managed false + ip link set testbr up + timer=0 + ip addr show testbr + grep -q 'inet [1-9]' + let timer+=1 + '[' 1 -eq 30 ']' + sleep 1 + rc=0 + ip addr add 192.0.2.1/24 dev testbr + '[' 0 '!=' 0 ']' + ip -6 addr add 2001:DB8::1/32 dev testbr + '[' 0 '!=' 0 ']' + grep -q 'inet [1-9]' + ip addr show testbr + grep 'release 6' /etc/redhat-release + ip link set test1p master testbr + ip link set test2p master testbr + systemctl is-active firewalld inactive + dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces 13118 1727204074.04385: no more pending results, returning what we have 13118 1727204074.04389: results queue empty 13118 1727204074.04390: checking for any_errors_fatal 13118 1727204074.04396: done checking for any_errors_fatal 13118 1727204074.04396: checking for max_fail_percentage 13118 1727204074.04398: done checking for max_fail_percentage 13118 1727204074.04399: checking to see if all hosts have failed and the running result is not ok 13118 1727204074.04399: done checking to see if all hosts have failed 13118 1727204074.04400: getting the remaining hosts for this loop 13118 1727204074.04401: done getting the remaining hosts for this loop 13118 1727204074.04405: getting the next task for host managed-node2 13118 1727204074.04416: done getting next task for host managed-node2 13118 1727204074.04418: ^ task is: TASK: Include the task 'get_interface_stat.yml' 13118 1727204074.04421: ^ state is: HOST STATE: block=2, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204074.04424: getting variables 13118 1727204074.04425: in VariableManager get_vars() 13118 1727204074.04469: Calling all_inventory to load vars for managed-node2 13118 1727204074.04472: Calling groups_inventory to load vars for managed-node2 13118 1727204074.04476: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.04486: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.04488: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.04491: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.04682: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.04874: done with get_vars() 13118 1727204074.04884: done getting variables TASK [Include the task 'get_interface_stat.yml'] ******************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:3 Tuesday 24 September 2024 14:54:34 -0400 (0:00:01.864) 0:00:11.349 ***** 13118 1727204074.04967: entering _queue_task() for managed-node2/include_tasks 13118 1727204074.05210: worker is 1 (out of 1 available) 13118 1727204074.05222: exiting _queue_task() for managed-node2/include_tasks 13118 1727204074.05231: done queuing things up, now waiting for results queue to drain 13118 1727204074.05233: waiting for pending results... 13118 1727204074.05478: running TaskExecutor() for managed-node2/TASK: Include the task 'get_interface_stat.yml' 13118 1727204074.05583: in run() - task 0affcd87-79f5-56a3-0a64-000000000016 13118 1727204074.05600: variable 'ansible_search_path' from source: unknown 13118 1727204074.05606: variable 'ansible_search_path' from source: unknown 13118 1727204074.05646: calling self._execute() 13118 1727204074.05724: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.05733: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.05745: variable 'omit' from source: magic vars 13118 1727204074.06093: variable 'ansible_distribution_major_version' from source: facts 13118 1727204074.06113: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204074.06122: _execute() done 13118 1727204074.06127: dumping result to json 13118 1727204074.06134: done dumping result, returning 13118 1727204074.06141: done running TaskExecutor() for managed-node2/TASK: Include the task 'get_interface_stat.yml' [0affcd87-79f5-56a3-0a64-000000000016] 13118 1727204074.06150: sending task result for task 0affcd87-79f5-56a3-0a64-000000000016 13118 1727204074.06259: no more pending results, returning what we have 13118 1727204074.06265: in VariableManager get_vars() 13118 1727204074.06308: Calling all_inventory to load vars for managed-node2 13118 1727204074.06311: Calling groups_inventory to load vars for managed-node2 13118 1727204074.06313: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.06326: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.06329: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.06331: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.06506: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.06691: done with get_vars() 13118 1727204074.06700: variable 'ansible_search_path' from source: unknown 13118 1727204074.06701: variable 'ansible_search_path' from source: unknown 13118 1727204074.06717: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000016 13118 1727204074.06720: WORKER PROCESS EXITING 13118 1727204074.06748: we have included files to process 13118 1727204074.06749: generating all_blocks data 13118 1727204074.06750: done generating all_blocks data 13118 1727204074.06751: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13118 1727204074.06752: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13118 1727204074.06753: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13118 1727204074.06992: done processing included file 13118 1727204074.06994: iterating over new_blocks loaded from include file 13118 1727204074.06995: in VariableManager get_vars() 13118 1727204074.07008: done with get_vars() 13118 1727204074.07010: filtering new block on tags 13118 1727204074.07020: done filtering new block on tags 13118 1727204074.07021: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml for managed-node2 13118 1727204074.07025: extending task lists for all hosts with included blocks 13118 1727204074.07089: done extending task lists 13118 1727204074.07090: done processing included files 13118 1727204074.07091: results queue empty 13118 1727204074.07091: checking for any_errors_fatal 13118 1727204074.07095: done checking for any_errors_fatal 13118 1727204074.07096: checking for max_fail_percentage 13118 1727204074.07096: done checking for max_fail_percentage 13118 1727204074.07097: checking to see if all hosts have failed and the running result is not ok 13118 1727204074.07097: done checking to see if all hosts have failed 13118 1727204074.07098: getting the remaining hosts for this loop 13118 1727204074.07099: done getting the remaining hosts for this loop 13118 1727204074.07100: getting the next task for host managed-node2 13118 1727204074.07103: done getting next task for host managed-node2 13118 1727204074.07104: ^ task is: TASK: Get stat for interface {{ interface }} 13118 1727204074.07106: ^ state is: HOST STATE: block=2, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204074.07107: getting variables 13118 1727204074.07108: in VariableManager get_vars() 13118 1727204074.07116: Calling all_inventory to load vars for managed-node2 13118 1727204074.07118: Calling groups_inventory to load vars for managed-node2 13118 1727204074.07119: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.07122: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.07124: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.07125: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.07232: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.07340: done with get_vars() 13118 1727204074.07346: done getting variables 13118 1727204074.07461: variable 'interface' from source: task vars 13118 1727204074.07467: variable 'dhcp_interface1' from source: play vars 13118 1727204074.07515: variable 'dhcp_interface1' from source: play vars TASK [Get stat for interface test1] ******************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml:3 Tuesday 24 September 2024 14:54:34 -0400 (0:00:00.025) 0:00:11.375 ***** 13118 1727204074.07552: entering _queue_task() for managed-node2/stat 13118 1727204074.07760: worker is 1 (out of 1 available) 13118 1727204074.07775: exiting _queue_task() for managed-node2/stat 13118 1727204074.07788: done queuing things up, now waiting for results queue to drain 13118 1727204074.07789: waiting for pending results... 13118 1727204074.07944: running TaskExecutor() for managed-node2/TASK: Get stat for interface test1 13118 1727204074.08018: in run() - task 0affcd87-79f5-56a3-0a64-000000000152 13118 1727204074.08032: variable 'ansible_search_path' from source: unknown 13118 1727204074.08036: variable 'ansible_search_path' from source: unknown 13118 1727204074.08063: calling self._execute() 13118 1727204074.08123: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.08134: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.08144: variable 'omit' from source: magic vars 13118 1727204074.08401: variable 'ansible_distribution_major_version' from source: facts 13118 1727204074.08411: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204074.08417: variable 'omit' from source: magic vars 13118 1727204074.08455: variable 'omit' from source: magic vars 13118 1727204074.08533: variable 'interface' from source: task vars 13118 1727204074.08537: variable 'dhcp_interface1' from source: play vars 13118 1727204074.08583: variable 'dhcp_interface1' from source: play vars 13118 1727204074.08597: variable 'omit' from source: magic vars 13118 1727204074.08633: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204074.08660: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204074.08678: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204074.08692: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204074.08702: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204074.08726: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204074.08732: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.08734: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.08799: Set connection var ansible_timeout to 10 13118 1727204074.08809: Set connection var ansible_pipelining to False 13118 1727204074.08812: Set connection var ansible_connection to ssh 13118 1727204074.08818: Set connection var ansible_shell_executable to /bin/sh 13118 1727204074.08822: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204074.08828: Set connection var ansible_shell_type to sh 13118 1727204074.08847: variable 'ansible_shell_executable' from source: unknown 13118 1727204074.08850: variable 'ansible_connection' from source: unknown 13118 1727204074.08852: variable 'ansible_module_compression' from source: unknown 13118 1727204074.08855: variable 'ansible_shell_type' from source: unknown 13118 1727204074.08857: variable 'ansible_shell_executable' from source: unknown 13118 1727204074.08859: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.08861: variable 'ansible_pipelining' from source: unknown 13118 1727204074.08866: variable 'ansible_timeout' from source: unknown 13118 1727204074.08871: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.09018: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204074.09026: variable 'omit' from source: magic vars 13118 1727204074.09034: starting attempt loop 13118 1727204074.09039: running the handler 13118 1727204074.09050: _low_level_execute_command(): starting 13118 1727204074.09057: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204074.09575: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.09597: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.09612: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.09624: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.09678: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.09689: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.09737: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.11357: stdout chunk (state=3): >>>/root <<< 13118 1727204074.11468: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204074.11532: stderr chunk (state=3): >>><<< 13118 1727204074.11536: stdout chunk (state=3): >>><<< 13118 1727204074.11555: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204074.11567: _low_level_execute_command(): starting 13118 1727204074.11573: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204074.1155424-14633-51699806564168 `" && echo ansible-tmp-1727204074.1155424-14633-51699806564168="` echo /root/.ansible/tmp/ansible-tmp-1727204074.1155424-14633-51699806564168 `" ) && sleep 0' 13118 1727204074.12305: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.12309: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.12348: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204074.12362: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.12372: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.12377: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.12446: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.12451: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204074.12472: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.12537: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.14403: stdout chunk (state=3): >>>ansible-tmp-1727204074.1155424-14633-51699806564168=/root/.ansible/tmp/ansible-tmp-1727204074.1155424-14633-51699806564168 <<< 13118 1727204074.14505: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204074.14601: stderr chunk (state=3): >>><<< 13118 1727204074.14621: stdout chunk (state=3): >>><<< 13118 1727204074.14961: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204074.1155424-14633-51699806564168=/root/.ansible/tmp/ansible-tmp-1727204074.1155424-14633-51699806564168 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204074.14970: variable 'ansible_module_compression' from source: unknown 13118 1727204074.14973: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13118 1727204074.14975: variable 'ansible_facts' from source: unknown 13118 1727204074.14977: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204074.1155424-14633-51699806564168/AnsiballZ_stat.py 13118 1727204074.15052: Sending initial data 13118 1727204074.15055: Sent initial data (152 bytes) 13118 1727204074.16672: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.16675: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.16715: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.16718: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.16720: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.16797: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.16801: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204074.16803: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.16866: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.18601: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204074.18645: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204074.18686: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpl9ikyavl /root/.ansible/tmp/ansible-tmp-1727204074.1155424-14633-51699806564168/AnsiballZ_stat.py <<< 13118 1727204074.18718: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204074.19865: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204074.20059: stderr chunk (state=3): >>><<< 13118 1727204074.20063: stdout chunk (state=3): >>><<< 13118 1727204074.20088: done transferring module to remote 13118 1727204074.20091: _low_level_execute_command(): starting 13118 1727204074.20093: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204074.1155424-14633-51699806564168/ /root/.ansible/tmp/ansible-tmp-1727204074.1155424-14633-51699806564168/AnsiballZ_stat.py && sleep 0' 13118 1727204074.20837: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204074.20852: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.20874: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.20893: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.20938: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204074.20951: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204074.20972: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.20991: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204074.21003: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204074.21013: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204074.21024: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.21039: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.21053: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.21076: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204074.21102: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204074.21114: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.21185: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.21209: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204074.21224: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.21291: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.23134: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204074.23138: stdout chunk (state=3): >>><<< 13118 1727204074.23140: stderr chunk (state=3): >>><<< 13118 1727204074.23256: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204074.23260: _low_level_execute_command(): starting 13118 1727204074.23262: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204074.1155424-14633-51699806564168/AnsiballZ_stat.py && sleep 0' 13118 1727204074.24707: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204074.24714: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.24731: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.24742: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.24786: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204074.24889: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204074.24899: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.24913: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204074.24920: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204074.24927: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204074.24934: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.24943: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.24955: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.24962: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204074.24969: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204074.24979: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.25053: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.25072: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204074.25110: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.25216: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.38507: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/test1", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 25203, "dev": 21, "nlink": 1, "atime": 1727204072.77989, "mtime": 1727204072.77989, "ctime": 1727204072.77989, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/test1", "lnk_target": "../../devices/virtual/net/test1", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/test1", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13118 1727204074.39716: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204074.39721: stdout chunk (state=3): >>><<< 13118 1727204074.39723: stderr chunk (state=3): >>><<< 13118 1727204074.39899: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/test1", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 25203, "dev": 21, "nlink": 1, "atime": 1727204072.77989, "mtime": 1727204072.77989, "ctime": 1727204072.77989, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/test1", "lnk_target": "../../devices/virtual/net/test1", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/test1", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204074.39903: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/sys/class/net/test1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204074.1155424-14633-51699806564168/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204074.39910: _low_level_execute_command(): starting 13118 1727204074.39912: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204074.1155424-14633-51699806564168/ > /dev/null 2>&1 && sleep 0' 13118 1727204074.40878: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204074.40994: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.41011: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.41041: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.41091: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204074.41105: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204074.41119: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.41138: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204074.41150: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204074.41161: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204074.41176: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.41192: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.41214: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.41227: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204074.41238: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204074.41252: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.41335: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.41359: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204074.41380: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.41452: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.43401: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204074.43405: stdout chunk (state=3): >>><<< 13118 1727204074.43407: stderr chunk (state=3): >>><<< 13118 1727204074.43845: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204074.43849: handler run complete 13118 1727204074.43852: attempt loop complete, returning result 13118 1727204074.43854: _execute() done 13118 1727204074.43856: dumping result to json 13118 1727204074.43858: done dumping result, returning 13118 1727204074.43860: done running TaskExecutor() for managed-node2/TASK: Get stat for interface test1 [0affcd87-79f5-56a3-0a64-000000000152] 13118 1727204074.43862: sending task result for task 0affcd87-79f5-56a3-0a64-000000000152 ok: [managed-node2] => { "changed": false, "stat": { "atime": 1727204072.77989, "block_size": 4096, "blocks": 0, "ctime": 1727204072.77989, "dev": 21, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 25203, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": true, "isreg": false, "issock": false, "isuid": false, "lnk_source": "/sys/devices/virtual/net/test1", "lnk_target": "../../devices/virtual/net/test1", "mode": "0777", "mtime": 1727204072.77989, "nlink": 1, "path": "/sys/class/net/test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "wgrp": true, "woth": true, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } 13118 1727204074.44266: no more pending results, returning what we have 13118 1727204074.44270: results queue empty 13118 1727204074.44271: checking for any_errors_fatal 13118 1727204074.44272: done checking for any_errors_fatal 13118 1727204074.44273: checking for max_fail_percentage 13118 1727204074.44274: done checking for max_fail_percentage 13118 1727204074.44275: checking to see if all hosts have failed and the running result is not ok 13118 1727204074.44276: done checking to see if all hosts have failed 13118 1727204074.44277: getting the remaining hosts for this loop 13118 1727204074.44278: done getting the remaining hosts for this loop 13118 1727204074.44282: getting the next task for host managed-node2 13118 1727204074.44289: done getting next task for host managed-node2 13118 1727204074.44293: ^ task is: TASK: Assert that the interface is present - '{{ interface }}' 13118 1727204074.44296: ^ state is: HOST STATE: block=2, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204074.44300: getting variables 13118 1727204074.44301: in VariableManager get_vars() 13118 1727204074.44338: Calling all_inventory to load vars for managed-node2 13118 1727204074.44341: Calling groups_inventory to load vars for managed-node2 13118 1727204074.44343: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.44354: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.44356: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.44360: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.44545: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000152 13118 1727204074.44552: WORKER PROCESS EXITING 13118 1727204074.44568: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.44821: done with get_vars() 13118 1727204074.44836: done getting variables 13118 1727204074.45049: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=False, class_only=True) 13118 1727204074.45278: variable 'interface' from source: task vars 13118 1727204074.45282: variable 'dhcp_interface1' from source: play vars 13118 1727204074.45462: variable 'dhcp_interface1' from source: play vars TASK [Assert that the interface is present - 'test1'] ************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:5 Tuesday 24 September 2024 14:54:34 -0400 (0:00:00.379) 0:00:11.754 ***** 13118 1727204074.45497: entering _queue_task() for managed-node2/assert 13118 1727204074.45499: Creating lock for assert 13118 1727204074.45816: worker is 1 (out of 1 available) 13118 1727204074.45830: exiting _queue_task() for managed-node2/assert 13118 1727204074.45843: done queuing things up, now waiting for results queue to drain 13118 1727204074.45844: waiting for pending results... 13118 1727204074.46110: running TaskExecutor() for managed-node2/TASK: Assert that the interface is present - 'test1' 13118 1727204074.46236: in run() - task 0affcd87-79f5-56a3-0a64-000000000017 13118 1727204074.46254: variable 'ansible_search_path' from source: unknown 13118 1727204074.46261: variable 'ansible_search_path' from source: unknown 13118 1727204074.46304: calling self._execute() 13118 1727204074.46390: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.46405: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.46418: variable 'omit' from source: magic vars 13118 1727204074.46787: variable 'ansible_distribution_major_version' from source: facts 13118 1727204074.46803: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204074.46814: variable 'omit' from source: magic vars 13118 1727204074.46869: variable 'omit' from source: magic vars 13118 1727204074.46978: variable 'interface' from source: task vars 13118 1727204074.46992: variable 'dhcp_interface1' from source: play vars 13118 1727204074.47066: variable 'dhcp_interface1' from source: play vars 13118 1727204074.47089: variable 'omit' from source: magic vars 13118 1727204074.47138: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204074.47183: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204074.47211: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204074.47236: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204074.47251: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204074.47291: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204074.47300: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.47309: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.47422: Set connection var ansible_timeout to 10 13118 1727204074.47446: Set connection var ansible_pipelining to False 13118 1727204074.47454: Set connection var ansible_connection to ssh 13118 1727204074.47465: Set connection var ansible_shell_executable to /bin/sh 13118 1727204074.47477: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204074.47486: Set connection var ansible_shell_type to sh 13118 1727204074.47514: variable 'ansible_shell_executable' from source: unknown 13118 1727204074.47522: variable 'ansible_connection' from source: unknown 13118 1727204074.47532: variable 'ansible_module_compression' from source: unknown 13118 1727204074.47541: variable 'ansible_shell_type' from source: unknown 13118 1727204074.47546: variable 'ansible_shell_executable' from source: unknown 13118 1727204074.47551: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.47556: variable 'ansible_pipelining' from source: unknown 13118 1727204074.47561: variable 'ansible_timeout' from source: unknown 13118 1727204074.47568: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.47704: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204074.47720: variable 'omit' from source: magic vars 13118 1727204074.47733: starting attempt loop 13118 1727204074.47742: running the handler 13118 1727204074.47902: variable 'interface_stat' from source: set_fact 13118 1727204074.47935: Evaluated conditional (interface_stat.stat.exists): True 13118 1727204074.47946: handler run complete 13118 1727204074.47967: attempt loop complete, returning result 13118 1727204074.47978: _execute() done 13118 1727204074.47985: dumping result to json 13118 1727204074.47992: done dumping result, returning 13118 1727204074.48002: done running TaskExecutor() for managed-node2/TASK: Assert that the interface is present - 'test1' [0affcd87-79f5-56a3-0a64-000000000017] 13118 1727204074.48014: sending task result for task 0affcd87-79f5-56a3-0a64-000000000017 ok: [managed-node2] => { "changed": false } MSG: All assertions passed 13118 1727204074.48203: no more pending results, returning what we have 13118 1727204074.48208: results queue empty 13118 1727204074.48209: checking for any_errors_fatal 13118 1727204074.48223: done checking for any_errors_fatal 13118 1727204074.48224: checking for max_fail_percentage 13118 1727204074.48226: done checking for max_fail_percentage 13118 1727204074.48227: checking to see if all hosts have failed and the running result is not ok 13118 1727204074.48228: done checking to see if all hosts have failed 13118 1727204074.48231: getting the remaining hosts for this loop 13118 1727204074.48233: done getting the remaining hosts for this loop 13118 1727204074.48237: getting the next task for host managed-node2 13118 1727204074.48248: done getting next task for host managed-node2 13118 1727204074.48251: ^ task is: TASK: Include the task 'get_interface_stat.yml' 13118 1727204074.48254: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204074.48259: getting variables 13118 1727204074.48260: in VariableManager get_vars() 13118 1727204074.48310: Calling all_inventory to load vars for managed-node2 13118 1727204074.48313: Calling groups_inventory to load vars for managed-node2 13118 1727204074.48315: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.48328: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.48334: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.48338: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.48561: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.48795: done with get_vars() 13118 1727204074.48811: done getting variables 13118 1727204074.49120: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000017 13118 1727204074.49124: WORKER PROCESS EXITING TASK [Include the task 'get_interface_stat.yml'] ******************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:3 Tuesday 24 September 2024 14:54:34 -0400 (0:00:00.036) 0:00:11.791 ***** 13118 1727204074.49193: entering _queue_task() for managed-node2/include_tasks 13118 1727204074.49627: worker is 1 (out of 1 available) 13118 1727204074.49643: exiting _queue_task() for managed-node2/include_tasks 13118 1727204074.49656: done queuing things up, now waiting for results queue to drain 13118 1727204074.49657: waiting for pending results... 13118 1727204074.49925: running TaskExecutor() for managed-node2/TASK: Include the task 'get_interface_stat.yml' 13118 1727204074.50039: in run() - task 0affcd87-79f5-56a3-0a64-00000000001b 13118 1727204074.50058: variable 'ansible_search_path' from source: unknown 13118 1727204074.50068: variable 'ansible_search_path' from source: unknown 13118 1727204074.50110: calling self._execute() 13118 1727204074.50201: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.50215: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.50234: variable 'omit' from source: magic vars 13118 1727204074.50611: variable 'ansible_distribution_major_version' from source: facts 13118 1727204074.50627: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204074.50645: _execute() done 13118 1727204074.50653: dumping result to json 13118 1727204074.50666: done dumping result, returning 13118 1727204074.50676: done running TaskExecutor() for managed-node2/TASK: Include the task 'get_interface_stat.yml' [0affcd87-79f5-56a3-0a64-00000000001b] 13118 1727204074.50688: sending task result for task 0affcd87-79f5-56a3-0a64-00000000001b 13118 1727204074.50811: no more pending results, returning what we have 13118 1727204074.50817: in VariableManager get_vars() 13118 1727204074.50873: Calling all_inventory to load vars for managed-node2 13118 1727204074.50876: Calling groups_inventory to load vars for managed-node2 13118 1727204074.50879: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.50893: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.50896: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.50899: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.51157: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.51374: done with get_vars() 13118 1727204074.51386: variable 'ansible_search_path' from source: unknown 13118 1727204074.51387: variable 'ansible_search_path' from source: unknown 13118 1727204074.51427: we have included files to process 13118 1727204074.51431: generating all_blocks data 13118 1727204074.51433: done generating all_blocks data 13118 1727204074.51439: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13118 1727204074.51440: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13118 1727204074.51442: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13118 1727204074.51717: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000001b 13118 1727204074.51721: WORKER PROCESS EXITING 13118 1727204074.51938: done processing included file 13118 1727204074.51940: iterating over new_blocks loaded from include file 13118 1727204074.51941: in VariableManager get_vars() 13118 1727204074.51962: done with get_vars() 13118 1727204074.51967: filtering new block on tags 13118 1727204074.51986: done filtering new block on tags 13118 1727204074.51988: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml for managed-node2 13118 1727204074.51993: extending task lists for all hosts with included blocks 13118 1727204074.52115: done extending task lists 13118 1727204074.52116: done processing included files 13118 1727204074.52117: results queue empty 13118 1727204074.52118: checking for any_errors_fatal 13118 1727204074.52121: done checking for any_errors_fatal 13118 1727204074.52122: checking for max_fail_percentage 13118 1727204074.52123: done checking for max_fail_percentage 13118 1727204074.52124: checking to see if all hosts have failed and the running result is not ok 13118 1727204074.52125: done checking to see if all hosts have failed 13118 1727204074.52125: getting the remaining hosts for this loop 13118 1727204074.52126: done getting the remaining hosts for this loop 13118 1727204074.52131: getting the next task for host managed-node2 13118 1727204074.52136: done getting next task for host managed-node2 13118 1727204074.52138: ^ task is: TASK: Get stat for interface {{ interface }} 13118 1727204074.52140: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204074.52143: getting variables 13118 1727204074.52144: in VariableManager get_vars() 13118 1727204074.52161: Calling all_inventory to load vars for managed-node2 13118 1727204074.52164: Calling groups_inventory to load vars for managed-node2 13118 1727204074.52167: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.52172: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.52175: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.52177: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.52325: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.52541: done with get_vars() 13118 1727204074.52550: done getting variables 13118 1727204074.52723: variable 'interface' from source: task vars 13118 1727204074.52726: variable 'dhcp_interface2' from source: play vars 13118 1727204074.52792: variable 'dhcp_interface2' from source: play vars TASK [Get stat for interface test2] ******************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml:3 Tuesday 24 September 2024 14:54:34 -0400 (0:00:00.036) 0:00:11.828 ***** 13118 1727204074.52824: entering _queue_task() for managed-node2/stat 13118 1727204074.53102: worker is 1 (out of 1 available) 13118 1727204074.53117: exiting _queue_task() for managed-node2/stat 13118 1727204074.53137: done queuing things up, now waiting for results queue to drain 13118 1727204074.53139: waiting for pending results... 13118 1727204074.53452: running TaskExecutor() for managed-node2/TASK: Get stat for interface test2 13118 1727204074.53625: in run() - task 0affcd87-79f5-56a3-0a64-00000000016a 13118 1727204074.53665: variable 'ansible_search_path' from source: unknown 13118 1727204074.53675: variable 'ansible_search_path' from source: unknown 13118 1727204074.53723: calling self._execute() 13118 1727204074.53816: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.53828: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.53845: variable 'omit' from source: magic vars 13118 1727204074.54310: variable 'ansible_distribution_major_version' from source: facts 13118 1727204074.54346: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204074.54362: variable 'omit' from source: magic vars 13118 1727204074.54422: variable 'omit' from source: magic vars 13118 1727204074.54603: variable 'interface' from source: task vars 13118 1727204074.54613: variable 'dhcp_interface2' from source: play vars 13118 1727204074.54692: variable 'dhcp_interface2' from source: play vars 13118 1727204074.54715: variable 'omit' from source: magic vars 13118 1727204074.54808: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204074.54851: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204074.54881: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204074.54914: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204074.54933: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204074.54971: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204074.54979: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.54986: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.55103: Set connection var ansible_timeout to 10 13118 1727204074.55132: Set connection var ansible_pipelining to False 13118 1727204074.55140: Set connection var ansible_connection to ssh 13118 1727204074.55149: Set connection var ansible_shell_executable to /bin/sh 13118 1727204074.55157: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204074.55165: Set connection var ansible_shell_type to sh 13118 1727204074.55190: variable 'ansible_shell_executable' from source: unknown 13118 1727204074.55197: variable 'ansible_connection' from source: unknown 13118 1727204074.55203: variable 'ansible_module_compression' from source: unknown 13118 1727204074.55210: variable 'ansible_shell_type' from source: unknown 13118 1727204074.55220: variable 'ansible_shell_executable' from source: unknown 13118 1727204074.55232: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.55241: variable 'ansible_pipelining' from source: unknown 13118 1727204074.55248: variable 'ansible_timeout' from source: unknown 13118 1727204074.55255: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.55562: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204074.55594: variable 'omit' from source: magic vars 13118 1727204074.55605: starting attempt loop 13118 1727204074.55615: running the handler 13118 1727204074.55674: _low_level_execute_command(): starting 13118 1727204074.55704: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204074.56332: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.56350: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204074.56366: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204074.56378: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.56423: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.56447: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.56483: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.58100: stdout chunk (state=3): >>>/root <<< 13118 1727204074.58207: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204074.58294: stderr chunk (state=3): >>><<< 13118 1727204074.58306: stdout chunk (state=3): >>><<< 13118 1727204074.58434: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204074.58438: _low_level_execute_command(): starting 13118 1727204074.58441: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204074.5833771-14715-269829206728013 `" && echo ansible-tmp-1727204074.5833771-14715-269829206728013="` echo /root/.ansible/tmp/ansible-tmp-1727204074.5833771-14715-269829206728013 `" ) && sleep 0' 13118 1727204074.58943: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204074.58946: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.58958: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.58995: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.59006: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.59008: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.59052: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.59070: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.59127: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.60999: stdout chunk (state=3): >>>ansible-tmp-1727204074.5833771-14715-269829206728013=/root/.ansible/tmp/ansible-tmp-1727204074.5833771-14715-269829206728013 <<< 13118 1727204074.61109: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204074.61203: stderr chunk (state=3): >>><<< 13118 1727204074.61217: stdout chunk (state=3): >>><<< 13118 1727204074.61271: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204074.5833771-14715-269829206728013=/root/.ansible/tmp/ansible-tmp-1727204074.5833771-14715-269829206728013 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204074.61598: variable 'ansible_module_compression' from source: unknown 13118 1727204074.61601: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13118 1727204074.61604: variable 'ansible_facts' from source: unknown 13118 1727204074.61606: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204074.5833771-14715-269829206728013/AnsiballZ_stat.py 13118 1727204074.61676: Sending initial data 13118 1727204074.61679: Sent initial data (153 bytes) 13118 1727204074.62678: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204074.62693: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.62708: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.62727: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.62773: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204074.62781: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204074.62791: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.62804: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204074.62813: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204074.62821: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204074.62826: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.62839: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.62851: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.62858: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204074.62866: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204074.62881: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.62955: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.62974: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204074.62985: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.63053: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.64850: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204074.64857: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204074.64906: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpyd11vrpe /root/.ansible/tmp/ansible-tmp-1727204074.5833771-14715-269829206728013/AnsiballZ_stat.py <<< 13118 1727204074.65219: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204074.66068: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204074.66279: stderr chunk (state=3): >>><<< 13118 1727204074.66282: stdout chunk (state=3): >>><<< 13118 1727204074.66284: done transferring module to remote 13118 1727204074.66287: _low_level_execute_command(): starting 13118 1727204074.66289: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204074.5833771-14715-269829206728013/ /root/.ansible/tmp/ansible-tmp-1727204074.5833771-14715-269829206728013/AnsiballZ_stat.py && sleep 0' 13118 1727204074.66916: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204074.66942: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.66958: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.66980: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.67025: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204074.67047: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204074.67063: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.67084: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204074.67097: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204074.67114: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204074.67127: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.67142: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.67170: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.67184: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204074.67196: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204074.67212: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.67299: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.67321: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204074.67338: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.67422: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.69232: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204074.69335: stderr chunk (state=3): >>><<< 13118 1727204074.69355: stdout chunk (state=3): >>><<< 13118 1727204074.69472: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204074.69475: _low_level_execute_command(): starting 13118 1727204074.69478: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204074.5833771-14715-269829206728013/AnsiballZ_stat.py && sleep 0' 13118 1727204074.70123: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204074.70143: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.70159: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.70180: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.70226: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204074.70247: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204074.70262: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.70283: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204074.70295: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204074.70307: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204074.70319: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204074.70332: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.70358: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.70374: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204074.70385: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204074.70398: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.70486: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.70509: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204074.70528: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.70614: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.83908: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/test2", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 25789, "dev": 21, "nlink": 1, "atime": 1727204072.7878604, "mtime": 1727204072.7878604, "ctime": 1727204072.7878604, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/test2", "lnk_target": "../../devices/virtual/net/test2", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/test2", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13118 1727204074.84885: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204074.84944: stderr chunk (state=3): >>><<< 13118 1727204074.84948: stdout chunk (state=3): >>><<< 13118 1727204074.84964: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/test2", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 25789, "dev": 21, "nlink": 1, "atime": 1727204072.7878604, "mtime": 1727204072.7878604, "ctime": 1727204072.7878604, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/test2", "lnk_target": "../../devices/virtual/net/test2", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/test2", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204074.85009: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/sys/class/net/test2', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204074.5833771-14715-269829206728013/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204074.85019: _low_level_execute_command(): starting 13118 1727204074.85024: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204074.5833771-14715-269829206728013/ > /dev/null 2>&1 && sleep 0' 13118 1727204074.85507: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.85513: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204074.85549: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204074.85563: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204074.85577: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204074.85620: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204074.85636: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204074.85649: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204074.85689: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204074.87516: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204074.87579: stderr chunk (state=3): >>><<< 13118 1727204074.87585: stdout chunk (state=3): >>><<< 13118 1727204074.87602: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204074.87610: handler run complete 13118 1727204074.87642: attempt loop complete, returning result 13118 1727204074.87645: _execute() done 13118 1727204074.87648: dumping result to json 13118 1727204074.87653: done dumping result, returning 13118 1727204074.87663: done running TaskExecutor() for managed-node2/TASK: Get stat for interface test2 [0affcd87-79f5-56a3-0a64-00000000016a] 13118 1727204074.87670: sending task result for task 0affcd87-79f5-56a3-0a64-00000000016a 13118 1727204074.87778: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000016a 13118 1727204074.87781: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "stat": { "atime": 1727204072.7878604, "block_size": 4096, "blocks": 0, "ctime": 1727204072.7878604, "dev": 21, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 25789, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": true, "isreg": false, "issock": false, "isuid": false, "lnk_source": "/sys/devices/virtual/net/test2", "lnk_target": "../../devices/virtual/net/test2", "mode": "0777", "mtime": 1727204072.7878604, "nlink": 1, "path": "/sys/class/net/test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "wgrp": true, "woth": true, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } 13118 1727204074.87893: no more pending results, returning what we have 13118 1727204074.87897: results queue empty 13118 1727204074.87898: checking for any_errors_fatal 13118 1727204074.87900: done checking for any_errors_fatal 13118 1727204074.87901: checking for max_fail_percentage 13118 1727204074.87903: done checking for max_fail_percentage 13118 1727204074.87904: checking to see if all hosts have failed and the running result is not ok 13118 1727204074.87905: done checking to see if all hosts have failed 13118 1727204074.87905: getting the remaining hosts for this loop 13118 1727204074.87906: done getting the remaining hosts for this loop 13118 1727204074.87910: getting the next task for host managed-node2 13118 1727204074.87917: done getting next task for host managed-node2 13118 1727204074.87919: ^ task is: TASK: Assert that the interface is present - '{{ interface }}' 13118 1727204074.87922: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204074.87926: getting variables 13118 1727204074.87927: in VariableManager get_vars() 13118 1727204074.87963: Calling all_inventory to load vars for managed-node2 13118 1727204074.87967: Calling groups_inventory to load vars for managed-node2 13118 1727204074.87969: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.87979: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.87980: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.87982: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.88133: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.88251: done with get_vars() 13118 1727204074.88260: done getting variables 13118 1727204074.88310: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204074.88399: variable 'interface' from source: task vars 13118 1727204074.88403: variable 'dhcp_interface2' from source: play vars 13118 1727204074.88450: variable 'dhcp_interface2' from source: play vars TASK [Assert that the interface is present - 'test2'] ************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:5 Tuesday 24 September 2024 14:54:34 -0400 (0:00:00.356) 0:00:12.184 ***** 13118 1727204074.88475: entering _queue_task() for managed-node2/assert 13118 1727204074.88675: worker is 1 (out of 1 available) 13118 1727204074.88690: exiting _queue_task() for managed-node2/assert 13118 1727204074.88702: done queuing things up, now waiting for results queue to drain 13118 1727204074.88703: waiting for pending results... 13118 1727204074.88858: running TaskExecutor() for managed-node2/TASK: Assert that the interface is present - 'test2' 13118 1727204074.88928: in run() - task 0affcd87-79f5-56a3-0a64-00000000001c 13118 1727204074.88943: variable 'ansible_search_path' from source: unknown 13118 1727204074.88946: variable 'ansible_search_path' from source: unknown 13118 1727204074.88980: calling self._execute() 13118 1727204074.89037: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.89044: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.89054: variable 'omit' from source: magic vars 13118 1727204074.89311: variable 'ansible_distribution_major_version' from source: facts 13118 1727204074.89321: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204074.89327: variable 'omit' from source: magic vars 13118 1727204074.89359: variable 'omit' from source: magic vars 13118 1727204074.89432: variable 'interface' from source: task vars 13118 1727204074.89436: variable 'dhcp_interface2' from source: play vars 13118 1727204074.89481: variable 'dhcp_interface2' from source: play vars 13118 1727204074.89500: variable 'omit' from source: magic vars 13118 1727204074.89540: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204074.89568: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204074.89584: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204074.89603: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204074.89614: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204074.89641: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204074.89644: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.89646: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.89717: Set connection var ansible_timeout to 10 13118 1727204074.89729: Set connection var ansible_pipelining to False 13118 1727204074.89732: Set connection var ansible_connection to ssh 13118 1727204074.89738: Set connection var ansible_shell_executable to /bin/sh 13118 1727204074.89743: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204074.89745: Set connection var ansible_shell_type to sh 13118 1727204074.89762: variable 'ansible_shell_executable' from source: unknown 13118 1727204074.89767: variable 'ansible_connection' from source: unknown 13118 1727204074.89770: variable 'ansible_module_compression' from source: unknown 13118 1727204074.89772: variable 'ansible_shell_type' from source: unknown 13118 1727204074.89775: variable 'ansible_shell_executable' from source: unknown 13118 1727204074.89777: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.89782: variable 'ansible_pipelining' from source: unknown 13118 1727204074.89784: variable 'ansible_timeout' from source: unknown 13118 1727204074.89788: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.89892: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204074.89900: variable 'omit' from source: magic vars 13118 1727204074.89905: starting attempt loop 13118 1727204074.89908: running the handler 13118 1727204074.90003: variable 'interface_stat' from source: set_fact 13118 1727204074.90019: Evaluated conditional (interface_stat.stat.exists): True 13118 1727204074.90022: handler run complete 13118 1727204074.90038: attempt loop complete, returning result 13118 1727204074.90041: _execute() done 13118 1727204074.90045: dumping result to json 13118 1727204074.90048: done dumping result, returning 13118 1727204074.90052: done running TaskExecutor() for managed-node2/TASK: Assert that the interface is present - 'test2' [0affcd87-79f5-56a3-0a64-00000000001c] 13118 1727204074.90061: sending task result for task 0affcd87-79f5-56a3-0a64-00000000001c 13118 1727204074.90144: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000001c 13118 1727204074.90147: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false } MSG: All assertions passed 13118 1727204074.90195: no more pending results, returning what we have 13118 1727204074.90198: results queue empty 13118 1727204074.90199: checking for any_errors_fatal 13118 1727204074.90207: done checking for any_errors_fatal 13118 1727204074.90208: checking for max_fail_percentage 13118 1727204074.90209: done checking for max_fail_percentage 13118 1727204074.90210: checking to see if all hosts have failed and the running result is not ok 13118 1727204074.90211: done checking to see if all hosts have failed 13118 1727204074.90211: getting the remaining hosts for this loop 13118 1727204074.90212: done getting the remaining hosts for this loop 13118 1727204074.90216: getting the next task for host managed-node2 13118 1727204074.90224: done getting next task for host managed-node2 13118 1727204074.90226: ^ task is: TASK: Backup the /etc/resolv.conf for initscript 13118 1727204074.90227: ^ state is: HOST STATE: block=2, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204074.90233: getting variables 13118 1727204074.90235: in VariableManager get_vars() 13118 1727204074.90277: Calling all_inventory to load vars for managed-node2 13118 1727204074.90279: Calling groups_inventory to load vars for managed-node2 13118 1727204074.90282: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.90291: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.90294: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.90296: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.90431: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.90548: done with get_vars() 13118 1727204074.90556: done getting variables 13118 1727204074.90601: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Backup the /etc/resolv.conf for initscript] ****************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml:28 Tuesday 24 September 2024 14:54:34 -0400 (0:00:00.021) 0:00:12.206 ***** 13118 1727204074.90621: entering _queue_task() for managed-node2/command 13118 1727204074.90820: worker is 1 (out of 1 available) 13118 1727204074.90836: exiting _queue_task() for managed-node2/command 13118 1727204074.90850: done queuing things up, now waiting for results queue to drain 13118 1727204074.90851: waiting for pending results... 13118 1727204074.91006: running TaskExecutor() for managed-node2/TASK: Backup the /etc/resolv.conf for initscript 13118 1727204074.91062: in run() - task 0affcd87-79f5-56a3-0a64-00000000001d 13118 1727204074.91074: variable 'ansible_search_path' from source: unknown 13118 1727204074.91102: calling self._execute() 13118 1727204074.91159: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.91163: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.91172: variable 'omit' from source: magic vars 13118 1727204074.91476: variable 'ansible_distribution_major_version' from source: facts 13118 1727204074.91485: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204074.91574: variable 'network_provider' from source: set_fact 13118 1727204074.91578: Evaluated conditional (network_provider == "initscripts"): False 13118 1727204074.91582: when evaluation is False, skipping this task 13118 1727204074.91584: _execute() done 13118 1727204074.91589: dumping result to json 13118 1727204074.91591: done dumping result, returning 13118 1727204074.91597: done running TaskExecutor() for managed-node2/TASK: Backup the /etc/resolv.conf for initscript [0affcd87-79f5-56a3-0a64-00000000001d] 13118 1727204074.91603: sending task result for task 0affcd87-79f5-56a3-0a64-00000000001d 13118 1727204074.91697: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000001d 13118 1727204074.91699: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "network_provider == \"initscripts\"", "skip_reason": "Conditional result was False" } 13118 1727204074.91771: no more pending results, returning what we have 13118 1727204074.91774: results queue empty 13118 1727204074.91775: checking for any_errors_fatal 13118 1727204074.91780: done checking for any_errors_fatal 13118 1727204074.91780: checking for max_fail_percentage 13118 1727204074.91787: done checking for max_fail_percentage 13118 1727204074.91789: checking to see if all hosts have failed and the running result is not ok 13118 1727204074.91789: done checking to see if all hosts have failed 13118 1727204074.91790: getting the remaining hosts for this loop 13118 1727204074.91791: done getting the remaining hosts for this loop 13118 1727204074.91794: getting the next task for host managed-node2 13118 1727204074.91799: done getting next task for host managed-node2 13118 1727204074.91801: ^ task is: TASK: TEST Add Bond with 2 ports 13118 1727204074.91803: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204074.91806: getting variables 13118 1727204074.91807: in VariableManager get_vars() 13118 1727204074.91837: Calling all_inventory to load vars for managed-node2 13118 1727204074.91839: Calling groups_inventory to load vars for managed-node2 13118 1727204074.91840: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.91847: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.91849: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.91850: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.92002: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.92118: done with get_vars() 13118 1727204074.92125: done getting variables 13118 1727204074.92170: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [TEST Add Bond with 2 ports] ********************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml:33 Tuesday 24 September 2024 14:54:34 -0400 (0:00:00.015) 0:00:12.221 ***** 13118 1727204074.92189: entering _queue_task() for managed-node2/debug 13118 1727204074.92374: worker is 1 (out of 1 available) 13118 1727204074.92386: exiting _queue_task() for managed-node2/debug 13118 1727204074.92398: done queuing things up, now waiting for results queue to drain 13118 1727204074.92400: waiting for pending results... 13118 1727204074.92580: running TaskExecutor() for managed-node2/TASK: TEST Add Bond with 2 ports 13118 1727204074.92634: in run() - task 0affcd87-79f5-56a3-0a64-00000000001e 13118 1727204074.92644: variable 'ansible_search_path' from source: unknown 13118 1727204074.92676: calling self._execute() 13118 1727204074.92736: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.92741: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.92750: variable 'omit' from source: magic vars 13118 1727204074.92999: variable 'ansible_distribution_major_version' from source: facts 13118 1727204074.93012: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204074.93017: variable 'omit' from source: magic vars 13118 1727204074.93035: variable 'omit' from source: magic vars 13118 1727204074.93063: variable 'omit' from source: magic vars 13118 1727204074.93100: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204074.93131: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204074.93145: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204074.93158: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204074.93170: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204074.93193: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204074.93196: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.93200: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.93268: Set connection var ansible_timeout to 10 13118 1727204074.93278: Set connection var ansible_pipelining to False 13118 1727204074.93280: Set connection var ansible_connection to ssh 13118 1727204074.93285: Set connection var ansible_shell_executable to /bin/sh 13118 1727204074.93290: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204074.93296: Set connection var ansible_shell_type to sh 13118 1727204074.93315: variable 'ansible_shell_executable' from source: unknown 13118 1727204074.93318: variable 'ansible_connection' from source: unknown 13118 1727204074.93322: variable 'ansible_module_compression' from source: unknown 13118 1727204074.93324: variable 'ansible_shell_type' from source: unknown 13118 1727204074.93327: variable 'ansible_shell_executable' from source: unknown 13118 1727204074.93332: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.93334: variable 'ansible_pipelining' from source: unknown 13118 1727204074.93336: variable 'ansible_timeout' from source: unknown 13118 1727204074.93338: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.93435: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204074.93456: variable 'omit' from source: magic vars 13118 1727204074.93478: starting attempt loop 13118 1727204074.93485: running the handler 13118 1727204074.93530: handler run complete 13118 1727204074.93550: attempt loop complete, returning result 13118 1727204074.93557: _execute() done 13118 1727204074.93563: dumping result to json 13118 1727204074.93573: done dumping result, returning 13118 1727204074.93601: done running TaskExecutor() for managed-node2/TASK: TEST Add Bond with 2 ports [0affcd87-79f5-56a3-0a64-00000000001e] 13118 1727204074.93620: sending task result for task 0affcd87-79f5-56a3-0a64-00000000001e ok: [managed-node2] => {} MSG: ################################################## 13118 1727204074.93776: no more pending results, returning what we have 13118 1727204074.93780: results queue empty 13118 1727204074.93781: checking for any_errors_fatal 13118 1727204074.93788: done checking for any_errors_fatal 13118 1727204074.93788: checking for max_fail_percentage 13118 1727204074.93790: done checking for max_fail_percentage 13118 1727204074.93791: checking to see if all hosts have failed and the running result is not ok 13118 1727204074.93791: done checking to see if all hosts have failed 13118 1727204074.93792: getting the remaining hosts for this loop 13118 1727204074.93793: done getting the remaining hosts for this loop 13118 1727204074.93797: getting the next task for host managed-node2 13118 1727204074.93804: done getting next task for host managed-node2 13118 1727204074.93809: ^ task is: TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role 13118 1727204074.93812: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204074.93823: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000001e 13118 1727204074.93826: WORKER PROCESS EXITING 13118 1727204074.93843: getting variables 13118 1727204074.93845: in VariableManager get_vars() 13118 1727204074.93886: Calling all_inventory to load vars for managed-node2 13118 1727204074.93889: Calling groups_inventory to load vars for managed-node2 13118 1727204074.93891: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.93899: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.93902: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.93905: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.94098: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.94315: done with get_vars() 13118 1727204074.94325: done getting variables TASK [fedora.linux_system_roles.network : Ensure ansible_facts used by role] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:4 Tuesday 24 September 2024 14:54:34 -0400 (0:00:00.022) 0:00:12.244 ***** 13118 1727204074.94428: entering _queue_task() for managed-node2/include_tasks 13118 1727204074.94803: worker is 1 (out of 1 available) 13118 1727204074.94818: exiting _queue_task() for managed-node2/include_tasks 13118 1727204074.94829: done queuing things up, now waiting for results queue to drain 13118 1727204074.94830: waiting for pending results... 13118 1727204074.94999: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role 13118 1727204074.95076: in run() - task 0affcd87-79f5-56a3-0a64-000000000026 13118 1727204074.95086: variable 'ansible_search_path' from source: unknown 13118 1727204074.95090: variable 'ansible_search_path' from source: unknown 13118 1727204074.95116: calling self._execute() 13118 1727204074.95173: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.95179: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.95186: variable 'omit' from source: magic vars 13118 1727204074.95482: variable 'ansible_distribution_major_version' from source: facts 13118 1727204074.95497: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204074.95502: _execute() done 13118 1727204074.95506: dumping result to json 13118 1727204074.95508: done dumping result, returning 13118 1727204074.95515: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role [0affcd87-79f5-56a3-0a64-000000000026] 13118 1727204074.95520: sending task result for task 0affcd87-79f5-56a3-0a64-000000000026 13118 1727204074.95603: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000026 13118 1727204074.95606: WORKER PROCESS EXITING 13118 1727204074.95643: no more pending results, returning what we have 13118 1727204074.95647: in VariableManager get_vars() 13118 1727204074.95690: Calling all_inventory to load vars for managed-node2 13118 1727204074.95692: Calling groups_inventory to load vars for managed-node2 13118 1727204074.95694: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.95703: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.95706: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.95708: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.95856: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.95971: done with get_vars() 13118 1727204074.95976: variable 'ansible_search_path' from source: unknown 13118 1727204074.95977: variable 'ansible_search_path' from source: unknown 13118 1727204074.96002: we have included files to process 13118 1727204074.96003: generating all_blocks data 13118 1727204074.96004: done generating all_blocks data 13118 1727204074.96008: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13118 1727204074.96009: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13118 1727204074.96010: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13118 1727204074.96466: done processing included file 13118 1727204074.96467: iterating over new_blocks loaded from include file 13118 1727204074.96469: in VariableManager get_vars() 13118 1727204074.96485: done with get_vars() 13118 1727204074.96486: filtering new block on tags 13118 1727204074.96497: done filtering new block on tags 13118 1727204074.96499: in VariableManager get_vars() 13118 1727204074.96512: done with get_vars() 13118 1727204074.96512: filtering new block on tags 13118 1727204074.96525: done filtering new block on tags 13118 1727204074.96526: in VariableManager get_vars() 13118 1727204074.96588: done with get_vars() 13118 1727204074.96590: filtering new block on tags 13118 1727204074.96607: done filtering new block on tags 13118 1727204074.96609: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml for managed-node2 13118 1727204074.96614: extending task lists for all hosts with included blocks 13118 1727204074.97585: done extending task lists 13118 1727204074.97586: done processing included files 13118 1727204074.97587: results queue empty 13118 1727204074.97588: checking for any_errors_fatal 13118 1727204074.97591: done checking for any_errors_fatal 13118 1727204074.97592: checking for max_fail_percentage 13118 1727204074.97593: done checking for max_fail_percentage 13118 1727204074.97593: checking to see if all hosts have failed and the running result is not ok 13118 1727204074.97594: done checking to see if all hosts have failed 13118 1727204074.97595: getting the remaining hosts for this loop 13118 1727204074.97596: done getting the remaining hosts for this loop 13118 1727204074.97598: getting the next task for host managed-node2 13118 1727204074.97603: done getting next task for host managed-node2 13118 1727204074.97605: ^ task is: TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present 13118 1727204074.97609: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204074.97617: getting variables 13118 1727204074.97618: in VariableManager get_vars() 13118 1727204074.97634: Calling all_inventory to load vars for managed-node2 13118 1727204074.97636: Calling groups_inventory to load vars for managed-node2 13118 1727204074.97638: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204074.97643: Calling all_plugins_play to load vars for managed-node2 13118 1727204074.97645: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204074.97648: Calling groups_plugins_play to load vars for managed-node2 13118 1727204074.97789: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204074.97988: done with get_vars() 13118 1727204074.97997: done getting variables TASK [fedora.linux_system_roles.network : Ensure ansible_facts used by role are present] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:3 Tuesday 24 September 2024 14:54:34 -0400 (0:00:00.036) 0:00:12.280 ***** 13118 1727204074.98074: entering _queue_task() for managed-node2/setup 13118 1727204074.98357: worker is 1 (out of 1 available) 13118 1727204074.98372: exiting _queue_task() for managed-node2/setup 13118 1727204074.98384: done queuing things up, now waiting for results queue to drain 13118 1727204074.98385: waiting for pending results... 13118 1727204074.98652: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present 13118 1727204074.98796: in run() - task 0affcd87-79f5-56a3-0a64-000000000188 13118 1727204074.98816: variable 'ansible_search_path' from source: unknown 13118 1727204074.98828: variable 'ansible_search_path' from source: unknown 13118 1727204074.98868: calling self._execute() 13118 1727204074.98944: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204074.98954: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204074.98968: variable 'omit' from source: magic vars 13118 1727204074.99318: variable 'ansible_distribution_major_version' from source: facts 13118 1727204074.99335: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204074.99551: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204075.01914: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204075.01996: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204075.02316: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204075.02355: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204075.02388: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204075.02474: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204075.02512: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204075.02544: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204075.02592: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204075.02612: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204075.02672: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204075.02699: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204075.02731: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204075.02778: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204075.02796: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204075.02960: variable '__network_required_facts' from source: role '' defaults 13118 1727204075.02976: variable 'ansible_facts' from source: unknown 13118 1727204075.03063: Evaluated conditional (__network_required_facts | difference(ansible_facts.keys() | list) | length > 0): False 13118 1727204075.03073: when evaluation is False, skipping this task 13118 1727204075.03080: _execute() done 13118 1727204075.03086: dumping result to json 13118 1727204075.03092: done dumping result, returning 13118 1727204075.03102: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present [0affcd87-79f5-56a3-0a64-000000000188] 13118 1727204075.03111: sending task result for task 0affcd87-79f5-56a3-0a64-000000000188 skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13118 1727204075.03251: no more pending results, returning what we have 13118 1727204075.03255: results queue empty 13118 1727204075.03256: checking for any_errors_fatal 13118 1727204075.03258: done checking for any_errors_fatal 13118 1727204075.03259: checking for max_fail_percentage 13118 1727204075.03260: done checking for max_fail_percentage 13118 1727204075.03261: checking to see if all hosts have failed and the running result is not ok 13118 1727204075.03262: done checking to see if all hosts have failed 13118 1727204075.03263: getting the remaining hosts for this loop 13118 1727204075.03266: done getting the remaining hosts for this loop 13118 1727204075.03270: getting the next task for host managed-node2 13118 1727204075.03280: done getting next task for host managed-node2 13118 1727204075.03284: ^ task is: TASK: fedora.linux_system_roles.network : Check if system is ostree 13118 1727204075.03289: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204075.03302: getting variables 13118 1727204075.03304: in VariableManager get_vars() 13118 1727204075.03346: Calling all_inventory to load vars for managed-node2 13118 1727204075.03350: Calling groups_inventory to load vars for managed-node2 13118 1727204075.03352: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204075.03362: Calling all_plugins_play to load vars for managed-node2 13118 1727204075.03367: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204075.03370: Calling groups_plugins_play to load vars for managed-node2 13118 1727204075.03583: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204075.03801: done with get_vars() 13118 1727204075.03813: done getting variables 13118 1727204075.04148: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000188 13118 1727204075.04151: WORKER PROCESS EXITING TASK [fedora.linux_system_roles.network : Check if system is ostree] *********** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:12 Tuesday 24 September 2024 14:54:35 -0400 (0:00:00.061) 0:00:12.341 ***** 13118 1727204075.04206: entering _queue_task() for managed-node2/stat 13118 1727204075.04458: worker is 1 (out of 1 available) 13118 1727204075.04474: exiting _queue_task() for managed-node2/stat 13118 1727204075.04487: done queuing things up, now waiting for results queue to drain 13118 1727204075.04488: waiting for pending results... 13118 1727204075.04791: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check if system is ostree 13118 1727204075.04973: in run() - task 0affcd87-79f5-56a3-0a64-00000000018a 13118 1727204075.05008: variable 'ansible_search_path' from source: unknown 13118 1727204075.05016: variable 'ansible_search_path' from source: unknown 13118 1727204075.05061: calling self._execute() 13118 1727204075.05147: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204075.05160: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204075.05176: variable 'omit' from source: magic vars 13118 1727204075.05670: variable 'ansible_distribution_major_version' from source: facts 13118 1727204075.05689: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204075.05861: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204075.06154: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204075.06204: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204075.06245: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204075.06283: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204075.06369: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204075.06397: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204075.06428: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204075.06465: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204075.06555: variable '__network_is_ostree' from source: set_fact 13118 1727204075.06573: Evaluated conditional (not __network_is_ostree is defined): False 13118 1727204075.06580: when evaluation is False, skipping this task 13118 1727204075.06585: _execute() done 13118 1727204075.06591: dumping result to json 13118 1727204075.06596: done dumping result, returning 13118 1727204075.06606: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check if system is ostree [0affcd87-79f5-56a3-0a64-00000000018a] 13118 1727204075.06615: sending task result for task 0affcd87-79f5-56a3-0a64-00000000018a skipping: [managed-node2] => { "changed": false, "false_condition": "not __network_is_ostree is defined", "skip_reason": "Conditional result was False" } 13118 1727204075.06762: no more pending results, returning what we have 13118 1727204075.06769: results queue empty 13118 1727204075.06770: checking for any_errors_fatal 13118 1727204075.06777: done checking for any_errors_fatal 13118 1727204075.06778: checking for max_fail_percentage 13118 1727204075.06779: done checking for max_fail_percentage 13118 1727204075.06780: checking to see if all hosts have failed and the running result is not ok 13118 1727204075.06781: done checking to see if all hosts have failed 13118 1727204075.06782: getting the remaining hosts for this loop 13118 1727204075.06783: done getting the remaining hosts for this loop 13118 1727204075.06787: getting the next task for host managed-node2 13118 1727204075.06796: done getting next task for host managed-node2 13118 1727204075.06799: ^ task is: TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree 13118 1727204075.06804: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204075.06819: getting variables 13118 1727204075.06821: in VariableManager get_vars() 13118 1727204075.06865: Calling all_inventory to load vars for managed-node2 13118 1727204075.06868: Calling groups_inventory to load vars for managed-node2 13118 1727204075.06871: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204075.06882: Calling all_plugins_play to load vars for managed-node2 13118 1727204075.06885: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204075.06888: Calling groups_plugins_play to load vars for managed-node2 13118 1727204075.07071: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204075.07289: done with get_vars() 13118 1727204075.07302: done getting variables 13118 1727204075.07366: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Set flag to indicate system is ostree] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:17 Tuesday 24 September 2024 14:54:35 -0400 (0:00:00.032) 0:00:12.374 ***** 13118 1727204075.07409: entering _queue_task() for managed-node2/set_fact 13118 1727204075.07432: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000018a 13118 1727204075.07442: WORKER PROCESS EXITING 13118 1727204075.07912: worker is 1 (out of 1 available) 13118 1727204075.07923: exiting _queue_task() for managed-node2/set_fact 13118 1727204075.07935: done queuing things up, now waiting for results queue to drain 13118 1727204075.07936: waiting for pending results... 13118 1727204075.08191: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree 13118 1727204075.08332: in run() - task 0affcd87-79f5-56a3-0a64-00000000018b 13118 1727204075.08350: variable 'ansible_search_path' from source: unknown 13118 1727204075.08356: variable 'ansible_search_path' from source: unknown 13118 1727204075.08399: calling self._execute() 13118 1727204075.08475: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204075.08490: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204075.08502: variable 'omit' from source: magic vars 13118 1727204075.08851: variable 'ansible_distribution_major_version' from source: facts 13118 1727204075.08870: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204075.09036: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204075.09380: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204075.09428: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204075.09470: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204075.09508: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204075.09599: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204075.09630: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204075.09661: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204075.09699: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204075.09798: variable '__network_is_ostree' from source: set_fact 13118 1727204075.09809: Evaluated conditional (not __network_is_ostree is defined): False 13118 1727204075.09816: when evaluation is False, skipping this task 13118 1727204075.09823: _execute() done 13118 1727204075.09829: dumping result to json 13118 1727204075.09837: done dumping result, returning 13118 1727204075.09847: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree [0affcd87-79f5-56a3-0a64-00000000018b] 13118 1727204075.09856: sending task result for task 0affcd87-79f5-56a3-0a64-00000000018b skipping: [managed-node2] => { "changed": false, "false_condition": "not __network_is_ostree is defined", "skip_reason": "Conditional result was False" } 13118 1727204075.09994: no more pending results, returning what we have 13118 1727204075.09999: results queue empty 13118 1727204075.10000: checking for any_errors_fatal 13118 1727204075.10005: done checking for any_errors_fatal 13118 1727204075.10006: checking for max_fail_percentage 13118 1727204075.10009: done checking for max_fail_percentage 13118 1727204075.10010: checking to see if all hosts have failed and the running result is not ok 13118 1727204075.10011: done checking to see if all hosts have failed 13118 1727204075.10012: getting the remaining hosts for this loop 13118 1727204075.10013: done getting the remaining hosts for this loop 13118 1727204075.10017: getting the next task for host managed-node2 13118 1727204075.10028: done getting next task for host managed-node2 13118 1727204075.10032: ^ task is: TASK: fedora.linux_system_roles.network : Check which services are running 13118 1727204075.10036: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204075.10052: getting variables 13118 1727204075.10054: in VariableManager get_vars() 13118 1727204075.10102: Calling all_inventory to load vars for managed-node2 13118 1727204075.10105: Calling groups_inventory to load vars for managed-node2 13118 1727204075.10107: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204075.10118: Calling all_plugins_play to load vars for managed-node2 13118 1727204075.10121: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204075.10124: Calling groups_plugins_play to load vars for managed-node2 13118 1727204075.10339: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204075.10556: done with get_vars() 13118 1727204075.10819: done getting variables 13118 1727204075.10852: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000018b 13118 1727204075.10855: WORKER PROCESS EXITING TASK [fedora.linux_system_roles.network : Check which services are running] **** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:21 Tuesday 24 September 2024 14:54:35 -0400 (0:00:00.035) 0:00:12.409 ***** 13118 1727204075.10929: entering _queue_task() for managed-node2/service_facts 13118 1727204075.10930: Creating lock for service_facts 13118 1727204075.11176: worker is 1 (out of 1 available) 13118 1727204075.11187: exiting _queue_task() for managed-node2/service_facts 13118 1727204075.11200: done queuing things up, now waiting for results queue to drain 13118 1727204075.11201: waiting for pending results... 13118 1727204075.11453: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check which services are running 13118 1727204075.11599: in run() - task 0affcd87-79f5-56a3-0a64-00000000018d 13118 1727204075.11618: variable 'ansible_search_path' from source: unknown 13118 1727204075.11624: variable 'ansible_search_path' from source: unknown 13118 1727204075.11666: calling self._execute() 13118 1727204075.11741: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204075.11756: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204075.11771: variable 'omit' from source: magic vars 13118 1727204075.12115: variable 'ansible_distribution_major_version' from source: facts 13118 1727204075.12131: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204075.12142: variable 'omit' from source: magic vars 13118 1727204075.12220: variable 'omit' from source: magic vars 13118 1727204075.12258: variable 'omit' from source: magic vars 13118 1727204075.12307: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204075.12346: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204075.12371: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204075.12394: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204075.12413: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204075.12444: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204075.12452: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204075.12459: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204075.12563: Set connection var ansible_timeout to 10 13118 1727204075.12583: Set connection var ansible_pipelining to False 13118 1727204075.12590: Set connection var ansible_connection to ssh 13118 1727204075.12598: Set connection var ansible_shell_executable to /bin/sh 13118 1727204075.12606: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204075.12616: Set connection var ansible_shell_type to sh 13118 1727204075.12642: variable 'ansible_shell_executable' from source: unknown 13118 1727204075.12650: variable 'ansible_connection' from source: unknown 13118 1727204075.12656: variable 'ansible_module_compression' from source: unknown 13118 1727204075.12661: variable 'ansible_shell_type' from source: unknown 13118 1727204075.12669: variable 'ansible_shell_executable' from source: unknown 13118 1727204075.12676: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204075.12682: variable 'ansible_pipelining' from source: unknown 13118 1727204075.12688: variable 'ansible_timeout' from source: unknown 13118 1727204075.12695: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204075.12891: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204075.12907: variable 'omit' from source: magic vars 13118 1727204075.12915: starting attempt loop 13118 1727204075.12922: running the handler 13118 1727204075.12944: _low_level_execute_command(): starting 13118 1727204075.12958: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204075.13696: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204075.13717: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204075.13734: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204075.13755: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204075.13804: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204075.13821: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204075.13836: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204075.13857: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204075.13894: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204075.13906: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204075.13923: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204075.13938: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204075.13955: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204075.13970: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204075.13982: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204075.13996: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204075.14103: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204075.14129: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204075.14150: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204075.14231: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204075.15891: stdout chunk (state=3): >>>/root <<< 13118 1727204075.15991: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204075.16068: stderr chunk (state=3): >>><<< 13118 1727204075.16071: stdout chunk (state=3): >>><<< 13118 1727204075.16083: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204075.16095: _low_level_execute_command(): starting 13118 1727204075.16104: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204075.1608412-14738-51610918467153 `" && echo ansible-tmp-1727204075.1608412-14738-51610918467153="` echo /root/.ansible/tmp/ansible-tmp-1727204075.1608412-14738-51610918467153 `" ) && sleep 0' 13118 1727204075.16566: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204075.16573: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204075.16621: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204075.16625: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204075.16627: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204075.16683: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204075.16686: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204075.16743: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204075.18622: stdout chunk (state=3): >>>ansible-tmp-1727204075.1608412-14738-51610918467153=/root/.ansible/tmp/ansible-tmp-1727204075.1608412-14738-51610918467153 <<< 13118 1727204075.18740: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204075.18792: stderr chunk (state=3): >>><<< 13118 1727204075.18797: stdout chunk (state=3): >>><<< 13118 1727204075.18817: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204075.1608412-14738-51610918467153=/root/.ansible/tmp/ansible-tmp-1727204075.1608412-14738-51610918467153 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204075.18858: variable 'ansible_module_compression' from source: unknown 13118 1727204075.18894: ANSIBALLZ: Using lock for service_facts 13118 1727204075.18897: ANSIBALLZ: Acquiring lock 13118 1727204075.18899: ANSIBALLZ: Lock acquired: 140051941515216 13118 1727204075.18902: ANSIBALLZ: Creating module 13118 1727204075.30632: ANSIBALLZ: Writing module into payload 13118 1727204075.30711: ANSIBALLZ: Writing module 13118 1727204075.30735: ANSIBALLZ: Renaming module 13118 1727204075.30738: ANSIBALLZ: Done creating module 13118 1727204075.30755: variable 'ansible_facts' from source: unknown 13118 1727204075.30802: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204075.1608412-14738-51610918467153/AnsiballZ_service_facts.py 13118 1727204075.30915: Sending initial data 13118 1727204075.30918: Sent initial data (161 bytes) 13118 1727204075.31633: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204075.31636: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204075.31672: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204075.31685: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204075.31740: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204075.31753: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204075.31803: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204075.33588: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204075.33623: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204075.33660: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmphpc3yhub /root/.ansible/tmp/ansible-tmp-1727204075.1608412-14738-51610918467153/AnsiballZ_service_facts.py <<< 13118 1727204075.33696: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204075.34777: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204075.34869: stderr chunk (state=3): >>><<< 13118 1727204075.34881: stdout chunk (state=3): >>><<< 13118 1727204075.34903: done transferring module to remote 13118 1727204075.34916: _low_level_execute_command(): starting 13118 1727204075.34920: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204075.1608412-14738-51610918467153/ /root/.ansible/tmp/ansible-tmp-1727204075.1608412-14738-51610918467153/AnsiballZ_service_facts.py && sleep 0' 13118 1727204075.35737: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204075.35743: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204075.35758: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204075.35797: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204075.35803: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204075.35822: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204075.35831: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204075.35910: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204075.35916: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204075.35934: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204075.35996: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204075.37766: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204075.37798: stderr chunk (state=3): >>><<< 13118 1727204075.37802: stdout chunk (state=3): >>><<< 13118 1727204075.37816: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204075.37819: _low_level_execute_command(): starting 13118 1727204075.37824: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204075.1608412-14738-51610918467153/AnsiballZ_service_facts.py && sleep 0' 13118 1727204075.38291: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204075.38295: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204075.38333: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204075.38336: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204075.38339: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204075.38394: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204075.38398: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204075.38404: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204075.38452: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204076.72954: stdout chunk (state=3): >>> {"ansible_facts": {"services": {"auditd.service": {"name": "auditd.service", "state": "running", "status": "enabled", "source": "systemd"}, "auth-rpcgss-module.service": {"name": "auth-rpcgss-module.service", "state": "stopped", "status": "static", "source": "systemd"}, "autofs.service": {"name": "autofs.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "chronyd.service": {"name": "chronyd.service", "state": "running", "status": "enabled", "source": "systemd"}, "cloud-config.service": {"name": "cloud-config.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-final.service": {"name": "cloud-final.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init-local.service": {"name": "cloud-init-local.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init.service": {"name": "cloud-init.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "crond.service": {"name": "crond.service", "state": "running", "status": "enabled", "source": "systemd"}, "dbus-broker.service": {"name": "dbus-broker.service", "state": "running", "status": "enabled", "source": "systemd"}, "display-manager.service": {"name": "display-manager.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "dnf-makecache.service": {"name": "dnf-makecache.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-cmdline.service": {"name": "dracut-cmdline.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-initqueue.service": {"name": "dracut-initqueue.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-mount.service": {"name": "dracut-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-mount.service": {"name": "dracut-pre-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-pivot.service": {"name": "dracut-pre-pivot.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-trigger.service": {"name": "dracut-pre-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-udev.service": {"name": "dracut-pre-udev.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown-onfailure.service": {"name": "dracut-shutdown-onfailure.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown.service": {"name": "dracut-shutdown.service", "state": "stopped", "status": "static", "source": "systemd"}, "emergency.service": {"name": "emergency.service", "state": "stopped", "status": "static", "source": "systemd"}, "getty@tty1.service": {"name": "getty@tty1.service", "state": "running", "status": "active", "source": "systemd"}, "gssproxy.service": {"name": "gssproxy.service", "state": "running", "status": "disabled", "source": "systemd"}, "hv_kvp_daemon.service": {"name": "hv_kvp_daemon.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "initrd-cleanup.service": {"name": "initrd-cleanup.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-parse-etc.service": {"name": "initrd-parse-etc.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-switch-root.service": {"name": "initrd-switch-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-udevadm-cleanup-db.service": {"name": "initrd-udevadm-cleanup-db.service", "state": "stopped", "status": "static", "source": "systemd"}, "irqbalance.service": {"name": "irqbalance.service", "state": "running", "status": "enabled", "source": "systemd"}, "kdump.service": {"name": "kdump.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "kmod-static-nodes.service": {"name": "kmod-static-nodes.service", "state": "stopped", "status": "static", "source": "systemd"}, "ldconfig.service": {"name": "ldconfig.service", "state": "stopped", "status": "static", "source": "systemd"}, "logrotate.service": {"name": "logrotate.service", "state": "stopped", "status": "static", "source": "systemd"}, "microcode.service": {"name": "microcode.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "modprobe@configfs.service": {"name": "modprobe@configfs.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@drm.service": {"name": "modprobe@drm.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@fuse.service": {"name": "modprobe@fuse.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "network.service": {"name": "network.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "NetworkManager-wait-online.service": {"name": "NetworkManager-wait-online.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "NetworkManager.service": {"name": "NetworkManager.service", "state": "running", "status": "enabled", "source": "systemd"}, "nfs-idmapd.service": {"name": "nfs-idmapd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-mountd.service": {"name": "nfs-mountd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-server.service": {"name": "nfs-server.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "nfs-utils.service": {"name": "nfs-utils.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfsdcld.service": {"name": "nfsdcld.service", "state": "stopped", "status": "static", "source": "systemd"}, "nis-domainname.service": {"name": "nis-domainname.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "ntpd.service": {"name": "ntpd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ntpdate.service": {"name": "ntpdate.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-quit-wait.service": {"name": "plymouth-quit-wait.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-start.service": {"name": "plymouth-start.service", "s<<< 13118 1727204076.72999: stdout chunk (state=3): >>>tate": "stopped", "status": "not-found", "source": "systemd"}, "rc-local.service": {"name": "rc-local.service", "state": "stopped", "status": "static", "source": "systemd"}, "rescue.service": {"name": "rescue.service", "state": "stopped", "status": "static", "source": "systemd"}, "restraintd.service": {"name": "restraintd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rngd.service": {"name": "rngd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rpc-gssd.service": {"name": "rpc-gssd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd-notify.service": {"name": "rpc-statd-notify.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd.service": {"name": "rpc-statd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-svcgssd.service": {"name": "rpc-svcgssd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rpcbind.service": {"name": "rpcbind.service", "state": "running", "status": "enabled", "source": "systemd"}, "rsyslog.service": {"name": "rsyslog.service", "state": "running", "status": "enabled", "source": "systemd"}, "selinux-autorelabel-mark.service": {"name": "selinux-autorelabel-mark.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "serial-getty@ttyS0.service": {"name": "serial-getty@ttyS0.service", "state": "running", "status": "active", "source": "systemd"}, "snapd.seeded.service": {"name": "snapd.seeded.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sntp.service": {"name": "sntp.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen.service": {"name": "sshd-keygen.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen@ecdsa.service": {"name": "sshd-keygen@ecdsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@ed25519.service": {"name": "sshd-keygen@ed25519.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@rsa.service": {"name": "sshd-keygen@rsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd.service": {"name": "sshd.service", "state": "running", "status": "enabled", "source": "systemd"}, "sssd-kcm.service": {"name": "sssd-kcm.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "sssd.service": {"name": "sssd.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "syslog.service": {"name": "syslog.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-ask-password-console.service": {"name": "systemd-ask-password-console.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-ask-password-wall.service": {"name": "systemd-ask-password-wall.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-binfmt.service": {"name": "systemd-binfmt.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-random-seed.service": {"name": "systemd-boot-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-update.service": {"name": "systemd-boot-update.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-firstboot.service": {"name": "systemd-firstboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-fsck-root.service": {"name": "systemd-fsck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hwdb-update.service": {"name": "systemd-hwdb-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-initctl.service": {"name": "systemd-initctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-catalog-update.service": {"name": "systemd-journal-catalog-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-flush.service": {"name": "systemd-journal-flush.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journald.service": {"name": "systemd-journald.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-logind.service": {"name": "systemd-logind.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-machine-id-commit.service": {"name": "systemd-machine-id-commit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-modules-load.service": {"name": "systemd-modules-load.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-network-generator.service": {"name": "systemd-network-generator.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-networkd-wait-online.service": {"name": "systemd-networkd-wait-online.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-pcrmachine.service": {"name": "systemd-pcrmachine.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-initrd.service": {"name": "systemd-pcrphase-initrd.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-sysinit.service": {"name": "systemd-pcrphase-sysinit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase.service": {"name": "systemd-pcrphase.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-random-seed.service": {"name": "systemd-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-remount-fs.service": {"name": "systemd-remount-fs.service", "state": "stopped", "status": "enabled-runtime", "source": "systemd"}, "systemd-repart.service": {"name": "systemd-repart.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-rfkill.service": {"name": "systemd-rfkill.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysctl.service": {"name": "systemd-sysctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysext.service": {"name": "systemd-sysext.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "systemd-sysusers.service": {"name": "systemd-sysusers.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-timesyncd.service": {"name": "systemd-timesyncd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-tmpfiles-clean.service": {"name": "systemd-tmpfiles-clean.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev.service": {"name": "systemd-tmpfiles-setup-dev.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup.service": {"name": "systemd-tmpfiles-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles.service": {"name": "systemd-tmpfiles.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-udev-settle.service": {"name": "systemd-udev-settle.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-trigger.service": {"name": "systemd-udev-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udevd.service": {"name": "systemd-udevd.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-update-done.service": {"name": "systemd-update-done.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp-runlevel.service": {"name": "systemd-update-utmp-runlevel.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp.service": {"name": "systemd-update-utmp.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-user-sessions.service": {"name": "systemd-user-sessions.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-vconsole-setup.service": {"name": "systemd-vconsole-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "user-runtim<<< 13118 1727204076.73006: stdout chunk (state=3): >>>e-dir@0.service": {"name": "user-runtime-dir@0.service", "state": "stopped", "status": "active", "source": "systemd"}, "user@0.service": {"name": "user@0.service", "state": "running", "status": "active", "source": "systemd"}, "wpa_supplicant.service": {"name": "wpa_supplicant.service", "state": "running", "status": "enabled", "source": "systemd"}, "ypbind.service": {"name": "ypbind.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "yppasswdd.service": {"name": "yppasswdd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ypserv.service": {"name": "ypserv.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ypxfrd.service": {"name": "ypxfrd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "autovt@.service": {"name": "autovt@.service", "state": "unknown", "status": "alias", "source": "systemd"}, "chrony-wait.service": {"name": "chrony-wait.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "chronyd-restricted.service": {"name": "chronyd-restricted.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "cloud-init-hotplugd.service": {"name": "cloud-init-hotplugd.service", "state": "inactive", "status": "static", "source": "systemd"}, "console-getty.service": {"name": "console-getty.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "container-getty@.service": {"name": "container-getty@.service", "state": "unknown", "status": "static", "source": "systemd"}, "cpupower.service": {"name": "cpupower.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dbus-org.freedesktop.hostname1.service": {"name": "dbus-org.freedesktop.hostname1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.locale1.service": {"name": "dbus-org.freedesktop.locale1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.login1.service": {"name": "dbus-org.freedesktop.login1.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.nm-dispatcher.service": {"name": "dbus-org.freedesktop.nm-dispatcher.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.timedate1.service": {"name": "dbus-org.freedesktop.timedate1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus.service": {"name": "dbus.service", "state": "active", "status": "alias", "source": "systemd"}, "debug-shell.service": {"name": "debug-shell.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnf-system-upgrade-cleanup.service": {"name": "dnf-system-upgrade-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "dnf-system-upgrade.service": {"name": "dnf-system-upgrade.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnsmasq.service": {"name": "dnsmasq.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "firewalld.service": {"name": "firewalld.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fstrim.service": {"name": "fstrim.service", "state": "inactive", "status": "static", "source": "systemd"}, "getty@.service": {"name": "getty@.service", "state": "unknown", "status": "enabled", "source": "systemd"}, "grub-boot-indeterminate.service": {"name": "grub-boot-indeterminate.service", "state": "inactive", "status": "static", "source": "systemd"}, "grub2-systemd-integration.service": {"name": "grub2-systemd-integration.service", "state": "inactive", "status": "static", "source": "systemd"}, "hostapd.service": {"name": "hostapd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "kvm_stat.service": {"name": "kvm_stat.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "man-db-cache-update.service": {"name": "man-db-cache-update.service", "state": "inactive", "status": "static", "source": "systemd"}, "man-db-restart-cache-update.service": {"name": "man-db-restart-cache-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "modprobe@.service": {"name": "modprobe@.service", "state": "unknown", "status": "static", "source": "systemd"}, "NetworkManager-dispatcher.service": {"name": "NetworkManager-dispatcher.service", "state": "active", "status": "enabled", "source": "systemd"}, "nfs-blkmap.service": {"name": "nfs-blkmap.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nftables.service": {"name": "nftables.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nm-priv-helper.service": {"name": "nm-priv-helper.service", "state": "inactive", "status": "static", "source": "systemd"}, "oddjobd.service": {"name": "oddjobd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "pam_namespace.service": {"name": "pam_namespace.service", "state": "inactive", "status": "static", "source": "systemd"}, "qemu-guest-agent.service": {"name": "qemu-guest-agent.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "quotaon.service": {"name": "quotaon.service", "state": "inactive", "status": "static", "source": "systemd"}, "rdisc.service": {"name": "rdisc.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "rpmdb-rebuild.service": {"name": "rpmdb-rebuild.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "selinux-autorelabel.service": {"name": "selinux-autorelabel.service", "state": "inactive", "status": "static", "source": "systemd"}, "selinux-check-proper-disable.service": {"name": "selinux-check-proper-disable.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "serial-getty@.service": {"name": "serial-getty@.service", "state": "unknown", "status": "indirect", "source": "systemd"}, "sshd-keygen@.service": {"name": "sshd-keygen@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "sshd@.service": {"name": "sshd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "sssd-autofs.service": {"name": "sssd-autofs.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-nss.service": {"name": "sssd-nss.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pac.service": {"name": "sssd-pac.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pam.service": {"name": "sssd-pam.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-ssh.service": {"name": "sssd-ssh.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-sudo.service": {"name": "sssd-sudo.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "system-update-cleanup.service": {"name": "system-update-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-backlight@.service": {"name": "systemd-backlight@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-bless-boot.service": {"name": "systemd-bless-boot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-boot-check-no-failures.service": {"name": "systemd-boot-check-no-failures.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-coredump@.service": {"name": "systemd-coredump@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-exit.service": {"name": "systemd-exit.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-fsck@.service": {"name": "systemd-fsck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-growfs-root.service": {"name": "systemd-growfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-growfs@.service": {"name": "systemd-growfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-halt.service": {"name": "systemd-halt.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hibernate-resume@.service": {"name": "systemd-hibernate-resume@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-hibernate.service": {"name": "systemd-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hostnamed.service": {"name": "systemd-hostnamed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hybrid-sleep.service": {"name": "systemd-hybrid-sleep.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-journald@.service": {"name": "systemd-journald@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-kexec.service": {"name": "systemd-kexec.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-localed.service": {"name": "systemd-localed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs-root.service": {"name": "systemd-pcrfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs@.service": {"name": "systemd-pcrfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-poweroff.service": {"name": "systemd-poweroff.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pstore.service": {"name": "systemd-pstore.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-quotacheck.service": {"name": "systemd-quotacheck.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-reboot.service": {"name": "systemd-reboot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend-then-hibernate.service": {"name": "systemd-suspend-then-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend.service": {"name": "systemd-suspend.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-sysupdate-reboot.service": {"name": "systemd-sysupdate-reboot.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-sysupdate.service": {"name": "systemd-sysupdate.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-timedated.service": {"name": "systemd-timedated.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-volatile-root.service": {"name": "systemd-volatile-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "teamd@.service": {"name": "teamd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user-runtime-dir@.service": {"name": "user-runtime-dir@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user@.service": {"name": "user@.service", "state": "unknown", "status": "static", "source": "systemd"}}}, "invocation": {"module_args": {}}} <<< 13118 1727204076.74381: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204076.74385: stdout chunk (state=3): >>><<< 13118 1727204076.74393: stderr chunk (state=3): >>><<< 13118 1727204076.74416: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"services": {"auditd.service": {"name": "auditd.service", "state": "running", "status": "enabled", "source": "systemd"}, "auth-rpcgss-module.service": {"name": "auth-rpcgss-module.service", "state": "stopped", "status": "static", "source": "systemd"}, "autofs.service": {"name": "autofs.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "chronyd.service": {"name": "chronyd.service", "state": "running", "status": "enabled", "source": "systemd"}, "cloud-config.service": {"name": "cloud-config.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-final.service": {"name": "cloud-final.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init-local.service": {"name": "cloud-init-local.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init.service": {"name": "cloud-init.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "crond.service": {"name": "crond.service", "state": "running", "status": "enabled", "source": "systemd"}, "dbus-broker.service": {"name": "dbus-broker.service", "state": "running", "status": "enabled", "source": "systemd"}, "display-manager.service": {"name": "display-manager.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "dnf-makecache.service": {"name": "dnf-makecache.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-cmdline.service": {"name": "dracut-cmdline.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-initqueue.service": {"name": "dracut-initqueue.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-mount.service": {"name": "dracut-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-mount.service": {"name": "dracut-pre-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-pivot.service": {"name": "dracut-pre-pivot.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-trigger.service": {"name": "dracut-pre-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-udev.service": {"name": "dracut-pre-udev.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown-onfailure.service": {"name": "dracut-shutdown-onfailure.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown.service": {"name": "dracut-shutdown.service", "state": "stopped", "status": "static", "source": "systemd"}, "emergency.service": {"name": "emergency.service", "state": "stopped", "status": "static", "source": "systemd"}, "getty@tty1.service": {"name": "getty@tty1.service", "state": "running", "status": "active", "source": "systemd"}, "gssproxy.service": {"name": "gssproxy.service", "state": "running", "status": "disabled", "source": "systemd"}, "hv_kvp_daemon.service": {"name": "hv_kvp_daemon.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "initrd-cleanup.service": {"name": "initrd-cleanup.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-parse-etc.service": {"name": "initrd-parse-etc.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-switch-root.service": {"name": "initrd-switch-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-udevadm-cleanup-db.service": {"name": "initrd-udevadm-cleanup-db.service", "state": "stopped", "status": "static", "source": "systemd"}, "irqbalance.service": {"name": "irqbalance.service", "state": "running", "status": "enabled", "source": "systemd"}, "kdump.service": {"name": "kdump.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "kmod-static-nodes.service": {"name": "kmod-static-nodes.service", "state": "stopped", "status": "static", "source": "systemd"}, "ldconfig.service": {"name": "ldconfig.service", "state": "stopped", "status": "static", "source": "systemd"}, "logrotate.service": {"name": "logrotate.service", "state": "stopped", "status": "static", "source": "systemd"}, "microcode.service": {"name": "microcode.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "modprobe@configfs.service": {"name": "modprobe@configfs.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@drm.service": {"name": "modprobe@drm.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@fuse.service": {"name": "modprobe@fuse.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "network.service": {"name": "network.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "NetworkManager-wait-online.service": {"name": "NetworkManager-wait-online.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "NetworkManager.service": {"name": "NetworkManager.service", "state": "running", "status": "enabled", "source": "systemd"}, "nfs-idmapd.service": {"name": "nfs-idmapd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-mountd.service": {"name": "nfs-mountd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-server.service": {"name": "nfs-server.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "nfs-utils.service": {"name": "nfs-utils.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfsdcld.service": {"name": "nfsdcld.service", "state": "stopped", "status": "static", "source": "systemd"}, "nis-domainname.service": {"name": "nis-domainname.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "ntpd.service": {"name": "ntpd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ntpdate.service": {"name": "ntpdate.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-quit-wait.service": {"name": "plymouth-quit-wait.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-start.service": {"name": "plymouth-start.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rc-local.service": {"name": "rc-local.service", "state": "stopped", "status": "static", "source": "systemd"}, "rescue.service": {"name": "rescue.service", "state": "stopped", "status": "static", "source": "systemd"}, "restraintd.service": {"name": "restraintd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rngd.service": {"name": "rngd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rpc-gssd.service": {"name": "rpc-gssd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd-notify.service": {"name": "rpc-statd-notify.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd.service": {"name": "rpc-statd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-svcgssd.service": {"name": "rpc-svcgssd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rpcbind.service": {"name": "rpcbind.service", "state": "running", "status": "enabled", "source": "systemd"}, "rsyslog.service": {"name": "rsyslog.service", "state": "running", "status": "enabled", "source": "systemd"}, "selinux-autorelabel-mark.service": {"name": "selinux-autorelabel-mark.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "serial-getty@ttyS0.service": {"name": "serial-getty@ttyS0.service", "state": "running", "status": "active", "source": "systemd"}, "snapd.seeded.service": {"name": "snapd.seeded.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sntp.service": {"name": "sntp.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen.service": {"name": "sshd-keygen.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen@ecdsa.service": {"name": "sshd-keygen@ecdsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@ed25519.service": {"name": "sshd-keygen@ed25519.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@rsa.service": {"name": "sshd-keygen@rsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd.service": {"name": "sshd.service", "state": "running", "status": "enabled", "source": "systemd"}, "sssd-kcm.service": {"name": "sssd-kcm.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "sssd.service": {"name": "sssd.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "syslog.service": {"name": "syslog.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-ask-password-console.service": {"name": "systemd-ask-password-console.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-ask-password-wall.service": {"name": "systemd-ask-password-wall.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-binfmt.service": {"name": "systemd-binfmt.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-random-seed.service": {"name": "systemd-boot-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-update.service": {"name": "systemd-boot-update.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-firstboot.service": {"name": "systemd-firstboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-fsck-root.service": {"name": "systemd-fsck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hwdb-update.service": {"name": "systemd-hwdb-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-initctl.service": {"name": "systemd-initctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-catalog-update.service": {"name": "systemd-journal-catalog-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-flush.service": {"name": "systemd-journal-flush.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journald.service": {"name": "systemd-journald.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-logind.service": {"name": "systemd-logind.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-machine-id-commit.service": {"name": "systemd-machine-id-commit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-modules-load.service": {"name": "systemd-modules-load.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-network-generator.service": {"name": "systemd-network-generator.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-networkd-wait-online.service": {"name": "systemd-networkd-wait-online.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-pcrmachine.service": {"name": "systemd-pcrmachine.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-initrd.service": {"name": "systemd-pcrphase-initrd.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-sysinit.service": {"name": "systemd-pcrphase-sysinit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase.service": {"name": "systemd-pcrphase.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-random-seed.service": {"name": "systemd-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-remount-fs.service": {"name": "systemd-remount-fs.service", "state": "stopped", "status": "enabled-runtime", "source": "systemd"}, "systemd-repart.service": {"name": "systemd-repart.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-rfkill.service": {"name": "systemd-rfkill.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysctl.service": {"name": "systemd-sysctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysext.service": {"name": "systemd-sysext.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "systemd-sysusers.service": {"name": "systemd-sysusers.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-timesyncd.service": {"name": "systemd-timesyncd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-tmpfiles-clean.service": {"name": "systemd-tmpfiles-clean.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev.service": {"name": "systemd-tmpfiles-setup-dev.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup.service": {"name": "systemd-tmpfiles-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles.service": {"name": "systemd-tmpfiles.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-udev-settle.service": {"name": "systemd-udev-settle.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-trigger.service": {"name": "systemd-udev-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udevd.service": {"name": "systemd-udevd.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-update-done.service": {"name": "systemd-update-done.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp-runlevel.service": {"name": "systemd-update-utmp-runlevel.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp.service": {"name": "systemd-update-utmp.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-user-sessions.service": {"name": "systemd-user-sessions.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-vconsole-setup.service": {"name": "systemd-vconsole-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "user-runtime-dir@0.service": {"name": "user-runtime-dir@0.service", "state": "stopped", "status": "active", "source": "systemd"}, "user@0.service": {"name": "user@0.service", "state": "running", "status": "active", "source": "systemd"}, "wpa_supplicant.service": {"name": "wpa_supplicant.service", "state": "running", "status": "enabled", "source": "systemd"}, "ypbind.service": {"name": "ypbind.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "yppasswdd.service": {"name": "yppasswdd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ypserv.service": {"name": "ypserv.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ypxfrd.service": {"name": "ypxfrd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "autovt@.service": {"name": "autovt@.service", "state": "unknown", "status": "alias", "source": "systemd"}, "chrony-wait.service": {"name": "chrony-wait.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "chronyd-restricted.service": {"name": "chronyd-restricted.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "cloud-init-hotplugd.service": {"name": "cloud-init-hotplugd.service", "state": "inactive", "status": "static", "source": "systemd"}, "console-getty.service": {"name": "console-getty.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "container-getty@.service": {"name": "container-getty@.service", "state": "unknown", "status": "static", "source": "systemd"}, "cpupower.service": {"name": "cpupower.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dbus-org.freedesktop.hostname1.service": {"name": "dbus-org.freedesktop.hostname1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.locale1.service": {"name": "dbus-org.freedesktop.locale1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.login1.service": {"name": "dbus-org.freedesktop.login1.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.nm-dispatcher.service": {"name": "dbus-org.freedesktop.nm-dispatcher.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.timedate1.service": {"name": "dbus-org.freedesktop.timedate1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus.service": {"name": "dbus.service", "state": "active", "status": "alias", "source": "systemd"}, "debug-shell.service": {"name": "debug-shell.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnf-system-upgrade-cleanup.service": {"name": "dnf-system-upgrade-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "dnf-system-upgrade.service": {"name": "dnf-system-upgrade.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnsmasq.service": {"name": "dnsmasq.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "firewalld.service": {"name": "firewalld.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fstrim.service": {"name": "fstrim.service", "state": "inactive", "status": "static", "source": "systemd"}, "getty@.service": {"name": "getty@.service", "state": "unknown", "status": "enabled", "source": "systemd"}, "grub-boot-indeterminate.service": {"name": "grub-boot-indeterminate.service", "state": "inactive", "status": "static", "source": "systemd"}, "grub2-systemd-integration.service": {"name": "grub2-systemd-integration.service", "state": "inactive", "status": "static", "source": "systemd"}, "hostapd.service": {"name": "hostapd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "kvm_stat.service": {"name": "kvm_stat.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "man-db-cache-update.service": {"name": "man-db-cache-update.service", "state": "inactive", "status": "static", "source": "systemd"}, "man-db-restart-cache-update.service": {"name": "man-db-restart-cache-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "modprobe@.service": {"name": "modprobe@.service", "state": "unknown", "status": "static", "source": "systemd"}, "NetworkManager-dispatcher.service": {"name": "NetworkManager-dispatcher.service", "state": "active", "status": "enabled", "source": "systemd"}, "nfs-blkmap.service": {"name": "nfs-blkmap.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nftables.service": {"name": "nftables.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nm-priv-helper.service": {"name": "nm-priv-helper.service", "state": "inactive", "status": "static", "source": "systemd"}, "oddjobd.service": {"name": "oddjobd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "pam_namespace.service": {"name": "pam_namespace.service", "state": "inactive", "status": "static", "source": "systemd"}, "qemu-guest-agent.service": {"name": "qemu-guest-agent.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "quotaon.service": {"name": "quotaon.service", "state": "inactive", "status": "static", "source": "systemd"}, "rdisc.service": {"name": "rdisc.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "rpmdb-rebuild.service": {"name": "rpmdb-rebuild.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "selinux-autorelabel.service": {"name": "selinux-autorelabel.service", "state": "inactive", "status": "static", "source": "systemd"}, "selinux-check-proper-disable.service": {"name": "selinux-check-proper-disable.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "serial-getty@.service": {"name": "serial-getty@.service", "state": "unknown", "status": "indirect", "source": "systemd"}, "sshd-keygen@.service": {"name": "sshd-keygen@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "sshd@.service": {"name": "sshd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "sssd-autofs.service": {"name": "sssd-autofs.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-nss.service": {"name": "sssd-nss.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pac.service": {"name": "sssd-pac.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pam.service": {"name": "sssd-pam.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-ssh.service": {"name": "sssd-ssh.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-sudo.service": {"name": "sssd-sudo.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "system-update-cleanup.service": {"name": "system-update-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-backlight@.service": {"name": "systemd-backlight@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-bless-boot.service": {"name": "systemd-bless-boot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-boot-check-no-failures.service": {"name": "systemd-boot-check-no-failures.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-coredump@.service": {"name": "systemd-coredump@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-exit.service": {"name": "systemd-exit.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-fsck@.service": {"name": "systemd-fsck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-growfs-root.service": {"name": "systemd-growfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-growfs@.service": {"name": "systemd-growfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-halt.service": {"name": "systemd-halt.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hibernate-resume@.service": {"name": "systemd-hibernate-resume@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-hibernate.service": {"name": "systemd-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hostnamed.service": {"name": "systemd-hostnamed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hybrid-sleep.service": {"name": "systemd-hybrid-sleep.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-journald@.service": {"name": "systemd-journald@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-kexec.service": {"name": "systemd-kexec.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-localed.service": {"name": "systemd-localed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs-root.service": {"name": "systemd-pcrfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs@.service": {"name": "systemd-pcrfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-poweroff.service": {"name": "systemd-poweroff.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pstore.service": {"name": "systemd-pstore.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-quotacheck.service": {"name": "systemd-quotacheck.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-reboot.service": {"name": "systemd-reboot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend-then-hibernate.service": {"name": "systemd-suspend-then-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend.service": {"name": "systemd-suspend.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-sysupdate-reboot.service": {"name": "systemd-sysupdate-reboot.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-sysupdate.service": {"name": "systemd-sysupdate.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-timedated.service": {"name": "systemd-timedated.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-volatile-root.service": {"name": "systemd-volatile-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "teamd@.service": {"name": "teamd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user-runtime-dir@.service": {"name": "user-runtime-dir@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user@.service": {"name": "user@.service", "state": "unknown", "status": "static", "source": "systemd"}}}, "invocation": {"module_args": {}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204076.75031: done with _execute_module (service_facts, {'_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'service_facts', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204075.1608412-14738-51610918467153/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204076.75037: _low_level_execute_command(): starting 13118 1727204076.75042: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204075.1608412-14738-51610918467153/ > /dev/null 2>&1 && sleep 0' 13118 1727204076.75746: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204076.75759: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204076.75771: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204076.75785: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204076.75824: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204076.75833: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204076.75840: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204076.75853: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204076.75865: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204076.75873: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204076.75881: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204076.75890: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204076.75901: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204076.75908: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204076.75914: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204076.75923: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204076.76001: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204076.76021: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204076.76034: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204076.76101: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204076.78709: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204076.78714: stdout chunk (state=3): >>><<< 13118 1727204076.78719: stderr chunk (state=3): >>><<< 13118 1727204076.78736: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204076.78744: handler run complete 13118 1727204076.78916: variable 'ansible_facts' from source: unknown 13118 1727204076.79050: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204076.79487: variable 'ansible_facts' from source: unknown 13118 1727204076.79609: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204076.79793: attempt loop complete, returning result 13118 1727204076.79796: _execute() done 13118 1727204076.79799: dumping result to json 13118 1727204076.79853: done dumping result, returning 13118 1727204076.79863: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check which services are running [0affcd87-79f5-56a3-0a64-00000000018d] 13118 1727204076.79870: sending task result for task 0affcd87-79f5-56a3-0a64-00000000018d ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13118 1727204076.80780: no more pending results, returning what we have 13118 1727204076.80784: results queue empty 13118 1727204076.80785: checking for any_errors_fatal 13118 1727204076.80788: done checking for any_errors_fatal 13118 1727204076.80789: checking for max_fail_percentage 13118 1727204076.80790: done checking for max_fail_percentage 13118 1727204076.80791: checking to see if all hosts have failed and the running result is not ok 13118 1727204076.80791: done checking to see if all hosts have failed 13118 1727204076.80792: getting the remaining hosts for this loop 13118 1727204076.80793: done getting the remaining hosts for this loop 13118 1727204076.80797: getting the next task for host managed-node2 13118 1727204076.80802: done getting next task for host managed-node2 13118 1727204076.80805: ^ task is: TASK: fedora.linux_system_roles.network : Check which packages are installed 13118 1727204076.80810: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204076.80818: getting variables 13118 1727204076.80819: in VariableManager get_vars() 13118 1727204076.80848: Calling all_inventory to load vars for managed-node2 13118 1727204076.80851: Calling groups_inventory to load vars for managed-node2 13118 1727204076.80853: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204076.80862: Calling all_plugins_play to load vars for managed-node2 13118 1727204076.80866: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204076.80869: Calling groups_plugins_play to load vars for managed-node2 13118 1727204076.81220: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204076.81719: done with get_vars() 13118 1727204076.81735: done getting variables 13118 1727204076.81772: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000018d 13118 1727204076.81775: WORKER PROCESS EXITING TASK [fedora.linux_system_roles.network : Check which packages are installed] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:26 Tuesday 24 September 2024 14:54:36 -0400 (0:00:01.709) 0:00:14.118 ***** 13118 1727204076.81852: entering _queue_task() for managed-node2/package_facts 13118 1727204076.81854: Creating lock for package_facts 13118 1727204076.82184: worker is 1 (out of 1 available) 13118 1727204076.82196: exiting _queue_task() for managed-node2/package_facts 13118 1727204076.82209: done queuing things up, now waiting for results queue to drain 13118 1727204076.82210: waiting for pending results... 13118 1727204076.82496: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check which packages are installed 13118 1727204076.82669: in run() - task 0affcd87-79f5-56a3-0a64-00000000018e 13118 1727204076.82694: variable 'ansible_search_path' from source: unknown 13118 1727204076.82706: variable 'ansible_search_path' from source: unknown 13118 1727204076.82752: calling self._execute() 13118 1727204076.82859: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204076.82877: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204076.82891: variable 'omit' from source: magic vars 13118 1727204076.83320: variable 'ansible_distribution_major_version' from source: facts 13118 1727204076.83343: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204076.83360: variable 'omit' from source: magic vars 13118 1727204076.83445: variable 'omit' from source: magic vars 13118 1727204076.83493: variable 'omit' from source: magic vars 13118 1727204076.83550: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204076.83600: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204076.83631: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204076.83658: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204076.83678: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204076.83716: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204076.83724: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204076.83735: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204076.83852: Set connection var ansible_timeout to 10 13118 1727204076.83875: Set connection var ansible_pipelining to False 13118 1727204076.83883: Set connection var ansible_connection to ssh 13118 1727204076.83893: Set connection var ansible_shell_executable to /bin/sh 13118 1727204076.83908: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204076.83916: Set connection var ansible_shell_type to sh 13118 1727204076.83946: variable 'ansible_shell_executable' from source: unknown 13118 1727204076.83954: variable 'ansible_connection' from source: unknown 13118 1727204076.83963: variable 'ansible_module_compression' from source: unknown 13118 1727204076.83975: variable 'ansible_shell_type' from source: unknown 13118 1727204076.83983: variable 'ansible_shell_executable' from source: unknown 13118 1727204076.83990: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204076.83998: variable 'ansible_pipelining' from source: unknown 13118 1727204076.84006: variable 'ansible_timeout' from source: unknown 13118 1727204076.84019: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204076.84247: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204076.84266: variable 'omit' from source: magic vars 13118 1727204076.84277: starting attempt loop 13118 1727204076.84285: running the handler 13118 1727204076.84306: _low_level_execute_command(): starting 13118 1727204076.84319: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204076.85173: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204076.85193: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204076.85210: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204076.85235: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204076.85282: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204076.85297: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204076.85311: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204076.85336: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204076.85349: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204076.85360: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204076.85375: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204076.85388: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204076.85407: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204076.85420: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204076.85435: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204076.85455: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204076.85541: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204076.85572: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204076.85590: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204076.85669: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204076.87321: stdout chunk (state=3): >>>/root <<< 13118 1727204076.87470: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204076.87522: stderr chunk (state=3): >>><<< 13118 1727204076.87525: stdout chunk (state=3): >>><<< 13118 1727204076.87554: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204076.87569: _low_level_execute_command(): starting 13118 1727204076.87576: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204076.8755333-14802-254494334316654 `" && echo ansible-tmp-1727204076.8755333-14802-254494334316654="` echo /root/.ansible/tmp/ansible-tmp-1727204076.8755333-14802-254494334316654 `" ) && sleep 0' 13118 1727204076.88232: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204076.88240: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204076.88249: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204076.88264: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204076.88312: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204076.88322: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204076.88334: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204076.88346: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204076.88352: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204076.88360: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204076.88369: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204076.88380: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204076.88389: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204076.88397: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204076.88403: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204076.88413: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204076.88497: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204076.88505: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204076.88508: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204076.88601: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204076.90457: stdout chunk (state=3): >>>ansible-tmp-1727204076.8755333-14802-254494334316654=/root/.ansible/tmp/ansible-tmp-1727204076.8755333-14802-254494334316654 <<< 13118 1727204076.90573: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204076.90666: stderr chunk (state=3): >>><<< 13118 1727204076.90670: stdout chunk (state=3): >>><<< 13118 1727204076.90689: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204076.8755333-14802-254494334316654=/root/.ansible/tmp/ansible-tmp-1727204076.8755333-14802-254494334316654 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204076.90736: variable 'ansible_module_compression' from source: unknown 13118 1727204076.90789: ANSIBALLZ: Using lock for package_facts 13118 1727204076.90792: ANSIBALLZ: Acquiring lock 13118 1727204076.90795: ANSIBALLZ: Lock acquired: 140051971228640 13118 1727204076.90797: ANSIBALLZ: Creating module 13118 1727204077.31151: ANSIBALLZ: Writing module into payload 13118 1727204077.31335: ANSIBALLZ: Writing module 13118 1727204077.31379: ANSIBALLZ: Renaming module 13118 1727204077.31382: ANSIBALLZ: Done creating module 13118 1727204077.31422: variable 'ansible_facts' from source: unknown 13118 1727204077.31635: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204076.8755333-14802-254494334316654/AnsiballZ_package_facts.py 13118 1727204077.31800: Sending initial data 13118 1727204077.31803: Sent initial data (162 bytes) 13118 1727204077.32897: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204077.32906: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204077.32917: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204077.32932: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204077.32969: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204077.32977: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204077.32987: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204077.33000: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204077.33008: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204077.33014: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204077.33022: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204077.33033: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204077.33043: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204077.33051: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204077.33060: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204077.33068: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204077.33151: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204077.33178: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204077.33183: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204077.33290: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204077.35126: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204077.35163: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204077.35211: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmp5esql4wf /root/.ansible/tmp/ansible-tmp-1727204076.8755333-14802-254494334316654/AnsiballZ_package_facts.py <<< 13118 1727204077.35241: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204077.37182: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204077.37302: stderr chunk (state=3): >>><<< 13118 1727204077.37306: stdout chunk (state=3): >>><<< 13118 1727204077.37309: done transferring module to remote 13118 1727204077.37311: _low_level_execute_command(): starting 13118 1727204077.37314: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204076.8755333-14802-254494334316654/ /root/.ansible/tmp/ansible-tmp-1727204076.8755333-14802-254494334316654/AnsiballZ_package_facts.py && sleep 0' 13118 1727204077.37907: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204077.37913: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204077.37923: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204077.37935: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204077.38158: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204077.38163: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204077.38166: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204077.38168: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204077.38170: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204077.38172: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204077.38176: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204077.38178: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204077.38180: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204077.38182: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204077.38185: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204077.38187: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204077.38189: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204077.38191: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204077.38221: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204077.38291: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204077.40153: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204077.40157: stdout chunk (state=3): >>><<< 13118 1727204077.40160: stderr chunk (state=3): >>><<< 13118 1727204077.40261: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204077.40266: _low_level_execute_command(): starting 13118 1727204077.40269: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204076.8755333-14802-254494334316654/AnsiballZ_package_facts.py && sleep 0' 13118 1727204077.40910: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204077.40940: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204077.40945: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204077.41011: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204077.41015: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204077.41023: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204077.41283: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204077.87542: stdout chunk (state=3): >>> {"ansible_facts": {"packages": {"libgcc": [{"name": "libgcc", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "linux-firmware-whence": [{"name": "linux-firmware-whence", "version": "20240905", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "tzdata": [{"name": "tzdata", "version": "2024a", "release": "2.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "linux-firmware": [{"name": "linux-firmware", "version": "20240905", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "gawk-all-langpacks": [{"name": "gawk-all-langpacks", "version": "5.1.0", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools-wheel": [{"name": "python3-setuptools-wheel", "version": "53.0.0", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "publicsuffix-list-dafsa": [{"name": "publicsuffix-list-dafsa", "version": "20210518", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "pcre2-syntax": [{"name": "pcre2-syntax", "version": "10.40", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "ncurses-base": [{"name": "ncurses-base", "version": "6.2", "release": "10.20210508.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libssh-config": [{"name": "libssh-config", "version": "0.10.4", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libreport-filesystem": [{"name": "libreport-filesystem", "version": "2.15.2", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf-data": [{"name": "dnf-data", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-misc": [{"name": "kbd-misc", "version": "2.4.0", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-legacy": [{"name": "kbd-legacy", "version": "2.4.0", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "hwdata": [{"name": "hwdata", "version": "0.348", "release": "9.15.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "fonts-filesystem": [{"name": "fonts-filesystem", "version": "2.0.5", "release": "7.el9.1", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dejavu-sans-fonts": [{"name": "dejavu-sans-fonts", "version": "2.37", "release": "18.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-core-font-en": [{"name": "langpacks-core-font-en", "version": "3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "firewalld-filesystem": [{"name": "firewalld-filesystem", "version": "1.3.4", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "coreutils-common": [{"name": "coreutils-common", "version": "8.32", "release": "36.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "centos-gpg-keys": [{"name": "centos-gpg-keys", "version": "9.0", "release": "26.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-repos": [{"name": "centos-stream-repos", "version": "9.0", "release": "26.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-release": [{"name": "centos-stream-release", "version": "9.0", "release": "26.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "setup": [{"name": "setup", "version": "2.13.7", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "filesystem": [{"name": "filesystem", "version": "3.16", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "basesystem": [{"name": "basesystem", "version": "11", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "glibc-gconv-extra": [{"name": "glibc-gconv-extra", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-langpack-en": [{"name": "glibc-langpack-en", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-common": [{"name": "glibc-common", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gli<<< 13118 1727204077.87576: stdout chunk (state=3): >>>bc": [{"name": "glibc", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses-libs": [{"name": "ncurses-libs", "version": "6.2", "release": "10.20210508.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bash": [{"name": "bash", "version": "5.1.8", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib": [{"name": "zlib", "version": "1.2.11", "release": "41.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-libs": [{"name": "xz-libs", "version": "5.2.5", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd": [{"name": "libzstd", "version": "1.5.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap": [{"name": "libcap", "version": "2.48", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "popt": [{"name": "popt", "version": "1.18", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bzip2-libs": [{"name": "bzip2-libs", "version": "1.0.8", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt": [{"name": "libxcrypt", "version": "4.4.18", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libuuid": [{"name": "libuuid", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sqlite-libs": [{"name": "sqlite-libs", "version": "3.34.1", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcom_err": [{"name": "libcom_err", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libstdc++": [{"name": "libstdc++", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmnl": [{"name": "libmnl", "version": "1.0.4", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf": [{"name": "elfutils-libelf", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxml2": [{"name": "libxml2", "version": "2.9.13", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies": [{"name": "crypto-policies", "version": "20240828", "release": "2.git626aa59.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "readline": [{"name": "readline", "version": "8.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "alternatives": [{"name": "alternatives", "version": "1.24", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng": [{"name": "libcap-ng", "version": "0.8.2", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit-libs": [{"name": "audit-libs", "version": "3.1.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libunistring": [{"name": "libunistring", "version": "0.9.10", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lua-libs": [{"name": "lua-libs", "version": "5.4.4", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libffi": [{"name": "libffi", "version": "3.4.2", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgpg-error": [{"name": "libgpg-error", "version": "1.42", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtalloc": [{"name": "libtalloc", "version": "2.4.2", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libidn2": [{"name": "libidn2", "version": "2.3.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gmp": [{"name": "gmp", "version": "6.2.0", "release": "13.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "keyutils-libs": [{"name": "keyutils-libs", "version": "1.6.3", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libattr": [{"name": "libattr", "version": "2.5.1", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libacl": [{"name": "libacl", "version": "2.3.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3": [{<<< 13118 1727204077.87581: stdout chunk (state=3): >>>"name": "libnl3", "version": "3.9.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsepol": [{"name": "libsepol", "version": "3.6", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsmartcols": [{"name": "libsmartcols", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lz4-libs": [{"name": "lz4-libs", "version": "1.9.3", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre2": [{"name": "pcre2", "version": "10.40", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux": [{"name": "libselinux", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sed": [{"name": "sed", "version": "4.8", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "findutils": [{"name": "findutils", "version": "4.8.0", "release": "7.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libsemanage": [{"name": "libsemanage", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "shadow-utils": [{"name": "shadow-utils", "version": "4.9", "release": "9.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libtevent": [{"name": "libtevent", "version": "0.16.1", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgcrypt": [{"name": "libgcrypt", "version": "1.10.0", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file-libs": [{"name": "file-libs", "version": "5.39", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libedit": [{"name": "libedit", "version": "3.1", "release": "38.20210216cvs.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "psmisc": [{"name": "psmisc", "version": "23.4", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "expat": [{"name": "expat", "version": "2.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm-libs": [{"name": "gdbm-libs", "version": "1.23", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "jansson": [{"name": "jansson", "version": "2.14", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "json-c": [{"name": "json-c", "version": "0.14", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtasn1": [{"name": "libtasn1", "version": "4.16.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit": [{"name": "p11-kit", "version": "0.25.3", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtdb": [{"name": "libtdb", "version": "1.4.10", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit-trust": [{"name": "p11-kit-trust", "version": "0.25.3", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file": [{"name": "file", "version": "5.39", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libassuan": [{"name": "libassuan", "version": "2.5.5", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbpf": [{"name": "libbpf", "version": "1.4.0", "release": "1.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libnftnl": [{"name": "libnftnl", "version": "1.2.6", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse-libs": [{"name": "fuse-libs", "version": "2.9.9", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbasicobjects": [{"name": "libbasicobjects", "version": "0.1.1", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcollection": [{"name": "libcollection", "version": "0.7.0", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdb": [{"name": "libdb", "version": "5.3.28", "release": "54.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute": [{"name": "iproute", "version": "6.2.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdhash": [{"name": "libdhash", "version": "0.5.0", "release": <<< 13118 1727204077.87591: stdout chunk (state=3): >>>"53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgomp": [{"name": "libgomp", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libref_array": [{"name": "libref_array", "version": "0.1.5", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libseccomp": [{"name": "libseccomp", "version": "2.5.2", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsigsegv": [{"name": "libsigsegv", "version": "2.13", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_idmap": [{"name": "libsss_idmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lzo": [{"name": "lzo", "version": "2.10", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "numactl-libs": [{"name": "numactl-libs", "version": "2.0.18", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre": [{"name": "pcre", "version": "8.44", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grep": [{"name": "grep", "version": "3.6", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssl-libs": [{"name": "openssl-libs", "version": "3.2.2", "release": "6.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "coreutils": [{"name": "coreutils", "version": "8.32", "release": "36.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ca-certificates": [{"name": "ca-certificates", "version": "2024.2.69_v8.0.303", "release": "91.4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "systemd-libs": [{"name": "systemd-libs", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libblkid": [{"name": "libblkid", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-libs": [{"name": "dbus-libs", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libmount": [{"name": "libmount", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux-core": [{"name": "util-linux-core", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfdisk": [{"name": "libfdisk", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gzip": [{"name": "gzip", "version": "1.12", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod": [{"name": "kmod", "version": "28", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod-libs": [{"name": "kmod-libs", "version": "28", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib": [{"name": "cracklib", "version": "2.9.6", "release": "27.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "which": [{"name": "which", "version": "2.21", "release": "29.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib-dicts": [{"name": "cracklib-dicts", "version": "2.9.6", "release": "27.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-tools": [{"name": "dbus-tools", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "procps-ng": [{"name": "procps-ng", "version": "3.3.17", "release": "14.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssl": [{"name": "openssl", "version": "3.2.2", "release": "6.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libarchive": [{"name": "libarchive", "version": "3.5.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libevent": [{"name": "libevent", "version": "2.1.12", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_certmap": [{"name": "libsss_certmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz": [{"name": "xz", "version": "5.2.5", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "squashfs-tools": [{"name": "squashfs-tools", "version": "4.4"<<< 13118 1727204077.87617: stdout chunk (state=3): >>>, "release": "10.git1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcomps": [{"name": "libcomps", "version": "0.1.18", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libutempter": [{"name": "libutempter", "version": "1.2.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux-utils": [{"name": "libselinux-utils", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3-cli": [{"name": "libnl3-cli", "version": "3.9.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libteam": [{"name": "libteam", "version": "1.31", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "acl": [{"name": "acl", "version": "2.3.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-libs": [{"name": "gettext-libs", "version": "0.21", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext": [{"name": "gettext", "version": "0.21", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "attr": [{"name": "attr", "version": "2.5.1", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils": [{"name": "keyutils", "version": "1.6.3", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpfr": [{"name": "mpfr", "version": "4.1.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gawk": [{"name": "gawk", "version": "5.1.0", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpsl": [{"name": "libpsl", "version": "0.21.1", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libksba": [{"name": "libksba", "version": "1.5.1", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ethtool": [{"name": "ethtool", "version": "6.2", "release": "1.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "ipset-libs": [{"name": "ipset-libs", "version": "7.11", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ipset": [{"name": "ipset", "version": "7.11", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "groff-base": [{"name": "groff-base", "version": "1.22.4", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "snappy": [{"name": "snappy", "version": "1.1.8", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs-libs": [{"name": "e2fsprogs-libs", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libss": [{"name": "libss", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt-compat": [{"name": "libxcrypt-compat", "version": "4.4.18", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip-wheel": [{"name": "python3-pip-wheel", "version": "21.3.1", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python-unversioned-command": [{"name": "python-unversioned-command", "version": "3.9.19", "release": "8.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3": [{"name": "python3", "version": "3.9.19", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libs": [{"name": "python3-libs", "version": "3.9.19", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libcomps": [{"name": "python3-libcomps", "version": "0.1.18", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-six": [{"name": "python3-six", "version": "1.15.0", "release": "9.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dateutil": [{"name": "python3-dateutil", "version": "2.8.1", "release": "7.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "python3-systemd": [{"name": "python3-systemd", "version": "234", "release": "19.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng-python3": [{"name": "libcap-ng-python3", "version": "0.8.2", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pigz": [{"name": "pigz", "version": "2.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "hostname": [{"name": "hostname", "version": "3.23", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools-libs": [{"name": "kernel-tools-libs", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "less": [{"name": "less", "version": "590", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-rpm-macros": [{"name": "systemd-rpm-macros", "version": "252", "release": "47.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "c-ares": [{"name": "c-ares", "version": "1.19.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpio": [{"name": "cpio", "version": "2.13", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "diffutils": [{"name": "diffutils", "version": "3.7", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "inih": [{"name": "inih", "version": "49", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbrotli": [{"name": "libbrotli", "version": "1.0.9", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcbor": [{"name": "libcbor", "version": "0.7.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdaemon": [{"name": "libdaemon", "version": "0.14", "release": "23.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "teamd": [{"name": "teamd", "version": "1.31", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libeconf": [{"name": "libeconf", "version": "0.4.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpwquality": [{"name": "libpwquality", "version": "1.4.4", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pam": [{"name": "pam", "version": "1.5.1", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux": [{"name": "util-linux", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus": [{"name": "dbus", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "systemd-pam": [{"name": "systemd-pam", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd": [{"name": "systemd", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-common": [{"name": "dbus-common", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dbus-broker": [{"name": "dbus-broker", "version": "28", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-common": [{"name": "grub2-common", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "cronie-anacron": [{"name": "cronie-anacron", "version": "1.5.7", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cronie": [{"name": "cronie", "version": "1.5.7", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crontabs": [{"name": "crontabs", "version": "1.11", "release": "26.20190603git.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "openssh": [{"name": "openssh", "version": "8.7p1", "release": "43.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc-modules": [{"name": "grub2-pc-modules", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "authselect-libs": [{"name": "authselect-libs", "version": "1.2.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-libs": [{"name": "device-mapper-libs", "version": "1.02.198", "release": "2.el9", "epoch": 9, "arch": "x86_64", "source": "rpm"}], "device-mapper": [{"name": "device-mapper", "version": "1.02.198", "release": "2.el9", "epoch": 9, "arch": "x86_64", "source": "rpm"}], "grub2-tools-minimal": [{"name": "grub2-tools-minimal", "version": "2.06", "rel<<< 13118 1727204077.87628: stdout chunk (state=3): >>>ease": "92.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "cryptsetup-libs": [{"name": "cryptsetup-libs", "version": "2.7.2", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kpartx": [{"name": "kpartx", "version": "0.8.7", "release": "32.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-default-yama-scope": [{"name": "elfutils-default-yama-scope", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "elfutils-libs": [{"name": "elfutils-libs", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-service": [{"name": "initscripts-service", "version": "10.11.8", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iputils": [{"name": "iputils", "version": "20210202", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi": [{"name": "libkcapi", "version": "1.4.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hmaccalc": [{"name": "libkcapi-hmaccalc", "version": "1.4.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "logrotate": [{"name": "logrotate", "version": "3.18.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oddjob": [{"name": "oddjob", "version": "0.34.7", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oddjob-mkhomedir": [{"name": "oddjob-mkhomedir", "version": "0.34.7", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect": [{"name": "authselect", "version": "1.2.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kbd": [{"name": "kbd", "version": "2.4.0", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libndp": [{"name": "libndp", "version": "1.9", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfnetlink": [{"name": "libnfnetlink", "version": "1.0.1", "release": "21.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnetfilter_conntrack": [{"name": "libnetfilter_conntrack", "version": "1.0.9", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-libs": [{"name": "iptables-libs", "version": "1.8.10", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-nft": [{"name": "iptables-nft", "version": "1.8.10", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nftables": [{"name": "nftables", "version": "1.0.9", "release": "3.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-nftables": [{"name": "python3-nftables", "version": "1.0.9", "release": "3.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libnghttp2": [{"name": "libnghttp2", "version": "1.43.0", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpath_utils": [{"name": "libpath_utils", "version": "0.2.1", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libini_config": [{"name": "libini_config", "version": "1.3.1", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpipeline": [{"name": "libpipeline", "version": "1.5.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_nss_idmap": [{"name": "libsss_nss_idmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_sudo": [{"name": "libsss_sudo", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto": [{"name": "libverto", "version": "0.3.2", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "krb5-libs": [{"name": "krb5-libs", "version": "1.21.1", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-lib": [{"name": "cyrus-sasl-lib", "version": "2.1.27", "release": "21.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openldap": [{"name": "openldap", "version": "2.6.6", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "<<< 13118 1727204077.87631: stdout chunk (state=3): >>>libssh": [{"name": "libssh", "version": "0.10.4", "release": "13.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcurl": [{"name": "libcurl", "version": "7.76.1", "release": "31.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-debuginfod-client": [{"name": "elfutils-debuginfod-client", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils-gold": [{"name": "binutils-gold", "version": "2.35.2", "release": "54.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils": [{"name": "binutils", "version": "2.35.2", "release": "54.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tss": [{"name": "tpm2-tss", "version": "3.2.3", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-udev": [{"name": "systemd-udev", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut": [{"name": "dracut", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-modules-core": [{"name": "kernel-modules-core", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-core": [{"name": "kernel-core", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-modules": [{"name": "kernel-modules", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-squash": [{"name": "dracut-squash", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfido2": [{"name": "libfido2", "version": "1.13.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "os-prober": [{"name": "os-prober", "version": "1.77", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools": [{"name": "grub2-tools", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "ima-evm-utils": [{"name": "ima-evm-utils", "version": "1.5", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "curl": [{"name": "curl", "version": "7.76.1", "release": "31.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm": [{"name": "rpm", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-libs": [{"name": "rpm-libs", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grubby": [{"name": "grubby", "version": "8.40", "release": "63.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsolv": [{"name": "libsolv", "version": "0.7.24", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "policycoreutils": [{"name": "policycoreutils", "version": "3.6", "release": "2.1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "selinux-policy": [{"name": "selinux-policy", "version": "38.1.45", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "selinux-policy-targeted": [{"name": "selinux-policy-targeted", "version": "38.1.45", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpm-build-libs": [{"name": "rpm-build-libs", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-systemd-inhibit": [{"name": "rpm-plugin-systemd-inhibit", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-client": [{"name": "sssd-client", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libyaml": [{"name": "libyaml", "version": "0.2.5", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lmdb-libs": [{"name": "lmdb-libs", "version": "0.9.29", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libldb": [{"name": "libldb", "version": "2.9.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-common": [{"name": "sssd-common", "version": "2.9.5", "rel<<< 13118 1727204077.87660: stdout chunk (state=3): >>>ease": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nettle": [{"name": "nettle", "version": "3.9.1", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnutls": [{"name": "gnutls", "version": "3.8.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glib2": [{"name": "glib2", "version": "2.68.4", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dbus": [{"name": "python3-dbus", "version": "1.2.18", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-libnm": [{"name": "NetworkManager-libnm", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "NetworkManager": [{"name": "NetworkManager", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libmodulemd": [{"name": "libmodulemd", "version": "2.13.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gobject-introspection": [{"name": "gobject-introspection", "version": "1.68.0", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gobject-base-noarch": [{"name": "python3-gobject-base-noarch", "version": "3.40.1", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-gobject-base": [{"name": "python3-gobject-base", "version": "3.40.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-firewall": [{"name": "python3-firewall", "version": "1.3.4", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libuser": [{"name": "libuser", "version": "0.63", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "npth": [{"name": "npth", "version": "1.6", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2": [{"name": "gnupg2", "version": "2.3.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gpgme": [{"name": "gpgme", "version": "1.15.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "librepo": [{"name": "librepo", "version": "1.14.5", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdnf": [{"name": "libdnf", "version": "0.69.0", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libdnf": [{"name": "python3-libdnf", "version": "0.69.0", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-hawkey": [{"name": "python3-hawkey", "version": "0.69.0", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gpg": [{"name": "python3-gpg", "version": "1.15.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sign-libs": [{"name": "rpm-sign-libs", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpm": [{"name": "python3-rpm", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dnf": [{"name": "python3-dnf", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf": [{"name": "dnf", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dnf-plugins-core": [{"name": "python3-dnf-plugins-core", "version": "4.3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "oniguruma": [{"name": "oniguruma", "version": "6.9.6", "release": "1.el9.6", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jq": [{"name": "jq", "version": "1.6", "release": "17.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-network": [{"name": "dracut-network", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pciutils-libs": [{"name": "pciutils-libs", "version": "3.7.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sg3_utils-libs": [{"name": "sg3_utils-libs", "version": "1.47", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], <<< 13118 1727204077.87678: stdout chunk (state=3): >>>"slang": [{"name": "slang", "version": "2.3.2", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "newt": [{"name": "newt", "version": "0.52.21", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "userspace-rcu": [{"name": "userspace-rcu", "version": "0.12.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libestr": [{"name": "libestr", "version": "0.1.11", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfastjson": [{"name": "libfastjson", "version": "0.99.9", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsyslog-logrotate": [{"name": "rsyslog-logrotate", "version": "8.2310.0", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsyslog": [{"name": "rsyslog", "version": "8.2310.0", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "liburing": [{"name": "liburing", "version": "2.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "langpacks-core-en": [{"name": "langpacks-core-en", "version": "3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-en": [{"name": "langpacks-en", "version": "3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "qemu-guest-agent": [{"name": "qemu-guest-agent", "version": "9.0.0", "release": "10.el9", "epoch": 17, "arch": "x86_64", "source": "rpm"}], "xfsprogs": [{"name": "xfsprogs", "version": "6.4.0", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-tui": [{"name": "NetworkManager-tui", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "sg3_utils": [{"name": "sg3_utils", "version": "1.47", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools": [{"name": "kernel-tools", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kexec-tools": [{"name": "kexec-tools", "version": "2.0.27", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dnf-plugins-core": [{"name": "dnf-plugins-core", "version": "4.3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "yum": [{"name": "yum", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "passwd": [{"name": "passwd", "version": "0.80", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "firewalld": [{"name": "firewalld", "version": "1.3.4", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "NetworkManager-team": [{"name": "NetworkManager-team", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "initscripts-rename-device": [{"name": "initscripts-rename-device", "version": "10.11.8", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "irqbalance": [{"name": "irqbalance", "version": "1.9.4", "release": "1.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "chrony": [{"name": "chrony", "version": "4.6", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-kcm": [{"name": "sssd-kcm", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-selinux": [{"name": "rpm-plugin-selinux", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies-scripts": [{"name": "crypto-policies-scripts", "version": "20240828", "release": "2.git626aa59.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpm-plugin-audit": [{"name": "rpm-plugin-audit", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc": [{"name": "grub2-pc", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "openssh-clients": [{"name": "openssh-clients", "version": "8.7p1", "release": "43.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel": [{"name": "kernel", "version": "5.14.0", "r<<< 13118 1727204077.87716: stdout chunk (state=3): >>>elease": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-config-rescue": [{"name": "dracut-config-rescue", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo": [{"name": "sudo", "version": "1.9.5p2", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit": [{"name": "audit", "version": "3.1.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-server": [{"name": "openssh-server", "version": "8.7p1", "release": "43.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "man-db": [{"name": "man-db", "version": "2.9.3", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute-tc": [{"name": "iproute-tc", "version": "6.2.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect-compat": [{"name": "authselect-compat", "version": "1.2.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "parted": [{"name": "parted", "version": "3.5", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "microcode_ctl": [{"name": "microcode_ctl", "version": "20240531", "release": "1.el9", "epoch": 4, "arch": "noarch", "source": "rpm"}], "python3-libselinux": [{"name": "python3-libselinux", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs": [{"name": "e2fsprogs", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "prefixdevname": [{"name": "prefixdevname", "version": "0.1.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-minimal": [{"name": "vim-minimal", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "lshw": [{"name": "lshw", "version": "B.02.19.2", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses": [{"name": "ncurses", "version": "6.2", "release": "10.20210508.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsysfs": [{"name": "libsysfs", "version": "2.1.1", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lsscsi": [{"name": "lsscsi", "version": "0.32", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iwl100-firmware": [{"name": "iwl100-firmware", "version": "39.31.5.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl1000-firmware": [{"name": "iwl1000-firmware", "version": "39.31.5.1", "release": "146.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "iwl105-firmware": [{"name": "iwl105-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl135-firmware": [{"name": "iwl135-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl2000-firmware": [{"name": "iwl2000-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl2030-firmware": [{"name": "iwl2030-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl3160-firmware": [{"name": "iwl3160-firmware", "version": "25.30.13.0", "release": "146.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "iwl5000-firmware": [{"name": "iwl5000-firmware", "version": "8.83.5.1_1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl5150-firmware": [{"name": "iwl5150-firmware", "version": "8.24.2.2", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl6000g2a-firmware": [{"name": "iwl6000g2a-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl6050-firmware": [{"name": "iwl6050-firmware", "version": "41.28.5.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl7260-firmware": [{"name": "iwl7260-firmware", "version": "25.30.13.0", "release": "146.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "rootfiles"<<< 13118 1727204077.87722: stdout chunk (state=3): >>>: [{"name": "rootfiles", "version": "8.1", "release": "31.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "gpg-pubkey": [{"name": "gpg-pubkey", "version": "3228467c", "release": "613798eb", "epoch": null, "arch": null, "source": "rpm"}, {"name": "gpg-pubkey", "version": "8483c65d", "release": "5ccc5b19", "epoch": null, "arch": null, "source": "rpm"}], "epel-release": [{"name": "epel-release", "version": "9", "release": "8.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "nspr": [{"name": "nspr", "version": "4.35.0", "release": "14.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "boost-system": [{"name": "boost-system", "version": "1.75.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-util": [{"name": "nss-util", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "make": [{"name": "make", "version": "4.3", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "m4": [{"name": "m4", "version": "1.4.19", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmpc": [{"name": "libmpc", "version": "1.2.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "unzip": [{"name": "unzip", "version": "6.0", "release": "57.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "avahi-libs": [{"name": "avahi-libs", "version": "0.8", "release": "21.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zip": [{"name": "zip", "version": "3.0", "release": "35.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpp": [{"name": "cpp", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bison": [{"name": "bison", "version": "3.7.4", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "flex": [{"name": "flex", "version": "2.6.4", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-softokn-freebl": [{"name": "nss-softokn-freebl", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-softokn": [{"name": "nss-softokn", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss": [{"name": "nss", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-sysinit": [{"name": "nss-sysinit", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "boost-filesystem": [{"name": "boost-filesystem", "version": "1.75.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "boost-thread": [{"name": "boost-thread", "version": "1.75.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Digest": [{"name": "perl-Digest", "version": "1.19", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Digest-MD5": [{"name": "perl-Digest-MD5", "version": "2.58", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-B": [{"name": "perl-B", "version": "1.80", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-FileHandle": [{"name": "perl-FileHandle", "version": "2.03", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Data-Dumper": [{"name": "perl-Data-Dumper", "version": "2.174", "release": "462.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-libnet": [{"name": "perl-libnet", "version": "3.13", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-base": [{"name": "perl-base", "version": "2.27", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-URI": [{"name": "perl-URI", "version": "5.09", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-AutoLoader": [{"name": "perl-AutoLoader", "version": "5.74", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Mozilla-CA": [{"name": "perl-Mozilla-CA", "version": "20200520", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "pe<<< 13118 1727204077.87739: stdout chunk (state=3): >>>rl-if": [{"name": "perl-if", "version": "0.60.800", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-IP": [{"name": "perl-IO-Socket-IP", "version": "0.41", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Time-Local": [{"name": "perl-Time-Local", "version": "1.300", "release": "7.el9", "epoch": 2, "arch": "noarch", "source": "rpm"}], "perl-File-Path": [{"name": "perl-File-Path", "version": "2.18", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Escapes": [{"name": "perl-Pod-Escapes", "version": "1.07", "release": "460.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Text-Tabs+Wrap": [{"name": "perl-Text-Tabs+Wrap", "version": "2013.0523", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-SSL": [{"name": "perl-IO-Socket-SSL", "version": "2.073", "release": "2.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Net-SSLeay": [{"name": "perl-Net-SSLeay", "version": "1.94", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Class-Struct": [{"name": "perl-Class-Struct", "version": "0.66", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-POSIX": [{"name": "perl-POSIX", "version": "1.94", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Term-ANSIColor": [{"name": "perl-Term-ANSIColor", "version": "5.01", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-IPC-Open3": [{"name": "perl-IPC-Open3", "version": "1.21", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-subs": [{"name": "perl-subs", "version": "1.03", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-Temp": [{"name": "perl-File-Temp", "version": "0.231.100", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Term-Cap": [{"name": "perl-Term-Cap", "version": "1.17", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Simple": [{"name": "perl-Pod-Simple", "version": "3.42", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-HTTP-Tiny": [{"name": "perl-HTTP-Tiny", "version": "0.076", "release": "462.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Socket": [{"name": "perl-Socket", "version": "2.031", "release": "4.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-SelectSaver": [{"name": "perl-SelectSaver", "version": "1.02", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Symbol": [{"name": "perl-Symbol", "version": "1.08", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-stat": [{"name": "perl-File-stat", "version": "1.09", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-podlators": [{"name": "perl-podlators", "version": "4.14", "release": "460.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Pod-Perldoc": [{"name": "perl-Pod-Perldoc", "version": "3.28.01", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Fcntl": [{"name": "perl-Fcntl", "version": "1.13", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Text-ParseWords": [{"name": "perl-Text-ParseWords", "version": "3.30", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-mro": [{"name": "perl-mro", "version": "1.23", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IO": [{"name": "perl-IO", "version": "1.43", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-overloading": [{"name": "perl-overloading", "version": "0.02", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Pod-Usage": [{"name": "perl-Pod-Usage", "version": "2.01", "release": "4.el9", "epoch": 4, "arch": "noarch", "source": "rpm"}], "perl-Errno": [{"name": "perl-Errno", "version": "1.30", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}],<<< 13118 1727204077.87744: stdout chunk (state=3): >>> "perl-File-Basename": [{"name": "perl-File-Basename", "version": "2.85", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Std": [{"name": "perl-Getopt-Std", "version": "1.12", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-MIME-Base64": [{"name": "perl-MIME-Base64", "version": "3.16", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Scalar-List-Utils": [{"name": "perl-Scalar-List-Utils", "version": "1.56", "release": "462.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-constant": [{"name": "perl-constant", "version": "1.33", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Storable": [{"name": "perl-Storable", "version": "3.21", "release": "460.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "perl-overload": [{"name": "perl-overload", "version": "1.31", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-parent": [{"name": "perl-parent", "version": "0.238", "release": "460.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-vars": [{"name": "perl-vars", "version": "1.05", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Long": [{"name": "perl-Getopt-Long", "version": "2.52", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Carp": [{"name": "perl-Carp", "version": "1.50", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Exporter": [{"name": "perl-Exporter", "version": "5.74", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-NDBM_File": [{"name": "perl-NDBM_File", "version": "1.15", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-PathTools": [{"name": "perl-PathTools", "version": "3.78", "release": "461.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Encode": [{"name": "perl-Encode", "version": "3.08", "release": "462.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-libs": [{"name": "perl-libs", "version": "5.32.1", "release": "481.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-interpreter": [{"name": "perl-interpreter", "version": "5.32.1", "release": "481.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "aspell": [{"name": "aspell", "version": "0.60.8", "release": "8.el9", "epoch": 12, "arch": "x86_64", "source": "rpm"}], "tbb": [{"name": "tbb", "version": "2020.3", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dyninst": [{"name": "dyninst", "version": "12.1.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemtap-runtime": [{"name": "systemtap-runtime", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-headers": [{"name": "kernel-headers", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-headers": [{"name": "glibc-headers", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "strace": [{"name": "strace", "version": "5.18", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pkgconf-m4": [{"name": "pkgconf-m4", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libpkgconf": [{"name": "libpkgconf", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pkgconf": [{"name": "pkgconf", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pkgconf-pkg-config": [{"name": "pkgconf-pkg-config", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd-devel": [{"name": "libzstd-devel", "version": "1.5.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib-devel": [{"name": "zlib-devel", "version": "1.2.11", "release": "41.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf-devel": [{"name": "elfutils-libelf-devel", "version": "<<< 13118 1727204077.87751: stdout chunk (state=3): >>>0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-devel": [{"name": "glibc-devel", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt-devel": [{"name": "libxcrypt-devel", "version": "4.4.18", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gcc": [{"name": "gcc", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssl-devel": [{"name": "openssl-devel", "version": "3.2.2", "release": "6.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-devel": [{"name": "kernel-devel", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-devel": [{"name": "xz-devel", "version": "5.2.5", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-devel": [{"name": "elfutils-devel", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemtap-devel": [{"name": "systemtap-devel", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "efivar-libs": [{"name": "efivar-libs", "version": "38", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mokutil": [{"name": "mokutil", "version": "0.6.0", "release": "4.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "systemtap-client": [{"name": "systemtap-client", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemtap": [{"name": "systemtap", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "qa-tools": [{"name": "qa-tools", "version": "4.1", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libtirpc": [{"name": "libtirpc", "version": "1.3.3", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "git-core": [{"name": "git-core", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfsidmap": [{"name": "libnfsidmap", "version": "2.5.4", "release": "27.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "git-core-doc": [{"name": "git-core-doc", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpcbind": [{"name": "rpcbind", "version": "1.2.6", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wget": [{"name": "wget", "version": "1.21.1", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-lib": [{"name": "perl-lib", "version": "0.65", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Find": [{"name": "perl-File-Find", "version": "1.37", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Error": [{"name": "perl-Error", "version": "0.17029", "release": "7.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-DynaLoader": [{"name": "perl-DynaLoader", "version": "1.47", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-TermReadKey": [{"name": "perl-TermReadKey", "version": "2.38", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxslt": [{"name": "libxslt", "version": "1.1.34", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-lxml": [{"name": "python3-lxml", "version": "4.6.5", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gpm-libs": [{"name": "gpm-libs", "version": "1.20.7", "release": "29.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "emacs-filesystem": [{"name": "emacs-filesystem", "version": "27.2", "release": "10.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Git": [{"name": "perl-Git", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "git": [{"name": "git", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "yum-utils": [{"name": "yum-utils", "version": "4.3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "s<<< 13118 1727204077.87773: stdout chunk (state=3): >>>ource": "rpm"}], "vim-filesystem": [{"name": "vim-filesystem", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "noarch", "source": "rpm"}], "vim-common": [{"name": "vim-common", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "time": [{"name": "time", "version": "1.9", "release": "18.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tar": [{"name": "tar", "version": "1.34", "release": "7.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "quota-nls": [{"name": "quota-nls", "version": "4.09", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "quota": [{"name": "quota", "version": "4.09", "release": "4.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-pyyaml": [{"name": "python3-pyyaml", "version": "5.4.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libev": [{"name": "libev", "version": "4.33", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto-libev": [{"name": "libverto-libev", "version": "0.3.2", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gssproxy": [{"name": "gssproxy", "version": "0.8.4", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nfs-utils": [{"name": "nfs-utils", "version": "2.5.4", "release": "27.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "bc": [{"name": "bc", "version": "1.07.1", "release": "14.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "beakerlib-redhat": [{"name": "beakerlib-redhat", "version": "1", "release": "35.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "beakerlib": [{"name": "beakerlib", "version": "1.29.3", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "restraint": [{"name": "restraint", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "restraint-rhts": [{"name": "restraint-rhts", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-enhanced": [{"name": "vim-enhanced", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "sssd-nfs-idmap": [{"name": "sssd-nfs-idmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsync": [{"name": "rsync", "version": "3.2.3", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools": [{"name": "python3-setuptools", "version": "53.0.0", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-distro": [{"name": "python3-distro", "version": "1.5.0", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-idna": [{"name": "python3-idna", "version": "2.10", "release": "7.el9.1", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-setools": [{"name": "python3-setools", "version": "4.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pytz": [{"name": "python3-pytz", "version": "2021.1", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-babel": [{"name": "python3-babel", "version": "2.9.1", "release": "2.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyserial": [{"name": "python3-pyserial", "version": "3.4", "release": "12.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyrsistent": [{"name": "python3-pyrsistent", "version": "0.17.3", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-prettytable": [{"name": "python3-prettytable", "version": "0.7.2", "release": "27.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-oauthlib": [{"name": "python3-oauthlib", "version": "3.1.1", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-netifaces": [{"name": "python3-netifaces", "version": "0.10.6", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-markupsafe": [{"name": "python3-markupsafe", "version": "1.1.1", "release": "12.el<<< 13118 1727204077.87802: stdout chunk (state=3): >>>9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jinja2": [{"name": "python3-jinja2", "version": "2.11.3", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libsemanage": [{"name": "python3-libsemanage", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jsonpointer": [{"name": "python3-jsonpointer", "version": "2.0", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonpatch": [{"name": "python3-jsonpatch", "version": "1.21", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-configobj": [{"name": "python3-configobj", "version": "5.0.6", "release": "25.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-audit": [{"name": "python3-audit", "version": "3.1.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-attrs": [{"name": "python3-attrs", "version": "20.3.0", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema": [{"name": "python3-jsonschema", "version": "3.2.0", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libmaxminddb": [{"name": "libmaxminddb", "version": "1.5.2", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "geolite2-country": [{"name": "geolite2-country", "version": "20191217", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "geolite2-city": [{"name": "geolite2-city", "version": "20191217", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "ipcalc": [{"name": "ipcalc", "version": "1.0.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdisk": [{"name": "gdisk", "version": "1.0.7", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "checkpolicy": [{"name": "checkpolicy", "version": "3.6", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-policycoreutils": [{"name": "python3-policycoreutils", "version": "3.6", "release": "2.1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pysocks": [{"name": "python3-pysocks", "version": "1.7.1", "release": "12.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-urllib3": [{"name": "python3-urllib3", "version": "1.26.5", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-chardet": [{"name": "python3-chardet", "version": "4.0.0", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-requests": [{"name": "python3-requests", "version": "2.25.1", "release": "8.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "dhcp-common": [{"name": "dhcp-common", "version": "4.4.2", "release": "19.b1.el9", "epoch": 12, "arch": "noarch", "source": "rpm"}], "dhcp-client": [{"name": "dhcp-client", "version": "4.4.2", "release": "19.b1.el9", "epoch": 12, "arch": "x86_64", "source": "rpm"}], "cloud-init": [{"name": "cloud-init", "version": "23.4", "release": "19.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "cloud-utils-growpart": [{"name": "cloud-utils-growpart", "version": "0.33", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jitterentropy": [{"name": "jitterentropy", "version": "3.5.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rng-tools": [{"name": "rng-tools", "version": "6.16", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip": [{"name": "python3-pip", "version": "21.3.1", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "hostapd": [{"name": "hostapd", "version": "2.10", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wpa_supplicant": [{"name": "wpa_supplicant", "version": "2.10", "release": "5.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dnsmasq": [{"name": "dnsmasq", "version": "2.85", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}]}}, "invocation": {"module_args": {"manager": ["auto"], "strategy": "first"}}} <<< 13118 1727204077.89405: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204077.89409: stderr chunk (state=3): >>><<< 13118 1727204077.89414: stdout chunk (state=3): >>><<< 13118 1727204077.89466: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"packages": {"libgcc": [{"name": "libgcc", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "linux-firmware-whence": [{"name": "linux-firmware-whence", "version": "20240905", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "tzdata": [{"name": "tzdata", "version": "2024a", "release": "2.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "linux-firmware": [{"name": "linux-firmware", "version": "20240905", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "gawk-all-langpacks": [{"name": "gawk-all-langpacks", "version": "5.1.0", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools-wheel": [{"name": "python3-setuptools-wheel", "version": "53.0.0", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "publicsuffix-list-dafsa": [{"name": "publicsuffix-list-dafsa", "version": "20210518", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "pcre2-syntax": [{"name": "pcre2-syntax", "version": "10.40", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "ncurses-base": [{"name": "ncurses-base", "version": "6.2", "release": "10.20210508.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libssh-config": [{"name": "libssh-config", "version": "0.10.4", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libreport-filesystem": [{"name": "libreport-filesystem", "version": "2.15.2", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf-data": [{"name": "dnf-data", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-misc": [{"name": "kbd-misc", "version": "2.4.0", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-legacy": [{"name": "kbd-legacy", "version": "2.4.0", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "hwdata": [{"name": "hwdata", "version": "0.348", "release": "9.15.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "fonts-filesystem": [{"name": "fonts-filesystem", "version": "2.0.5", "release": "7.el9.1", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dejavu-sans-fonts": [{"name": "dejavu-sans-fonts", "version": "2.37", "release": "18.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-core-font-en": [{"name": "langpacks-core-font-en", "version": "3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "firewalld-filesystem": [{"name": "firewalld-filesystem", "version": "1.3.4", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "coreutils-common": [{"name": "coreutils-common", "version": "8.32", "release": "36.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "centos-gpg-keys": [{"name": "centos-gpg-keys", "version": "9.0", "release": "26.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-repos": [{"name": "centos-stream-repos", "version": "9.0", "release": "26.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-release": [{"name": "centos-stream-release", "version": "9.0", "release": "26.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "setup": [{"name": "setup", "version": "2.13.7", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "filesystem": [{"name": "filesystem", "version": "3.16", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "basesystem": [{"name": "basesystem", "version": "11", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "glibc-gconv-extra": [{"name": "glibc-gconv-extra", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-langpack-en": [{"name": "glibc-langpack-en", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-common": [{"name": "glibc-common", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc": [{"name": "glibc", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses-libs": [{"name": "ncurses-libs", "version": "6.2", "release": "10.20210508.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bash": [{"name": "bash", "version": "5.1.8", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib": [{"name": "zlib", "version": "1.2.11", "release": "41.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-libs": [{"name": "xz-libs", "version": "5.2.5", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd": [{"name": "libzstd", "version": "1.5.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap": [{"name": "libcap", "version": "2.48", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "popt": [{"name": "popt", "version": "1.18", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bzip2-libs": [{"name": "bzip2-libs", "version": "1.0.8", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt": [{"name": "libxcrypt", "version": "4.4.18", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libuuid": [{"name": "libuuid", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sqlite-libs": [{"name": "sqlite-libs", "version": "3.34.1", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcom_err": [{"name": "libcom_err", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libstdc++": [{"name": "libstdc++", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmnl": [{"name": "libmnl", "version": "1.0.4", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf": [{"name": "elfutils-libelf", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxml2": [{"name": "libxml2", "version": "2.9.13", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies": [{"name": "crypto-policies", "version": "20240828", "release": "2.git626aa59.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "readline": [{"name": "readline", "version": "8.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "alternatives": [{"name": "alternatives", "version": "1.24", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng": [{"name": "libcap-ng", "version": "0.8.2", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit-libs": [{"name": "audit-libs", "version": "3.1.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libunistring": [{"name": "libunistring", "version": "0.9.10", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lua-libs": [{"name": "lua-libs", "version": "5.4.4", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libffi": [{"name": "libffi", "version": "3.4.2", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgpg-error": [{"name": "libgpg-error", "version": "1.42", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtalloc": [{"name": "libtalloc", "version": "2.4.2", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libidn2": [{"name": "libidn2", "version": "2.3.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gmp": [{"name": "gmp", "version": "6.2.0", "release": "13.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "keyutils-libs": [{"name": "keyutils-libs", "version": "1.6.3", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libattr": [{"name": "libattr", "version": "2.5.1", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libacl": [{"name": "libacl", "version": "2.3.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3": [{"name": "libnl3", "version": "3.9.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsepol": [{"name": "libsepol", "version": "3.6", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsmartcols": [{"name": "libsmartcols", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lz4-libs": [{"name": "lz4-libs", "version": "1.9.3", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre2": [{"name": "pcre2", "version": "10.40", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux": [{"name": "libselinux", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sed": [{"name": "sed", "version": "4.8", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "findutils": [{"name": "findutils", "version": "4.8.0", "release": "7.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libsemanage": [{"name": "libsemanage", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "shadow-utils": [{"name": "shadow-utils", "version": "4.9", "release": "9.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libtevent": [{"name": "libtevent", "version": "0.16.1", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgcrypt": [{"name": "libgcrypt", "version": "1.10.0", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file-libs": [{"name": "file-libs", "version": "5.39", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libedit": [{"name": "libedit", "version": "3.1", "release": "38.20210216cvs.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "psmisc": [{"name": "psmisc", "version": "23.4", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "expat": [{"name": "expat", "version": "2.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm-libs": [{"name": "gdbm-libs", "version": "1.23", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "jansson": [{"name": "jansson", "version": "2.14", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "json-c": [{"name": "json-c", "version": "0.14", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtasn1": [{"name": "libtasn1", "version": "4.16.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit": [{"name": "p11-kit", "version": "0.25.3", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtdb": [{"name": "libtdb", "version": "1.4.10", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit-trust": [{"name": "p11-kit-trust", "version": "0.25.3", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file": [{"name": "file", "version": "5.39", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libassuan": [{"name": "libassuan", "version": "2.5.5", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbpf": [{"name": "libbpf", "version": "1.4.0", "release": "1.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libnftnl": [{"name": "libnftnl", "version": "1.2.6", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse-libs": [{"name": "fuse-libs", "version": "2.9.9", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbasicobjects": [{"name": "libbasicobjects", "version": "0.1.1", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcollection": [{"name": "libcollection", "version": "0.7.0", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdb": [{"name": "libdb", "version": "5.3.28", "release": "54.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute": [{"name": "iproute", "version": "6.2.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdhash": [{"name": "libdhash", "version": "0.5.0", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgomp": [{"name": "libgomp", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libref_array": [{"name": "libref_array", "version": "0.1.5", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libseccomp": [{"name": "libseccomp", "version": "2.5.2", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsigsegv": [{"name": "libsigsegv", "version": "2.13", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_idmap": [{"name": "libsss_idmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lzo": [{"name": "lzo", "version": "2.10", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "numactl-libs": [{"name": "numactl-libs", "version": "2.0.18", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre": [{"name": "pcre", "version": "8.44", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grep": [{"name": "grep", "version": "3.6", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssl-libs": [{"name": "openssl-libs", "version": "3.2.2", "release": "6.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "coreutils": [{"name": "coreutils", "version": "8.32", "release": "36.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ca-certificates": [{"name": "ca-certificates", "version": "2024.2.69_v8.0.303", "release": "91.4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "systemd-libs": [{"name": "systemd-libs", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libblkid": [{"name": "libblkid", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-libs": [{"name": "dbus-libs", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libmount": [{"name": "libmount", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux-core": [{"name": "util-linux-core", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfdisk": [{"name": "libfdisk", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gzip": [{"name": "gzip", "version": "1.12", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod": [{"name": "kmod", "version": "28", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod-libs": [{"name": "kmod-libs", "version": "28", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib": [{"name": "cracklib", "version": "2.9.6", "release": "27.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "which": [{"name": "which", "version": "2.21", "release": "29.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib-dicts": [{"name": "cracklib-dicts", "version": "2.9.6", "release": "27.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-tools": [{"name": "dbus-tools", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "procps-ng": [{"name": "procps-ng", "version": "3.3.17", "release": "14.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssl": [{"name": "openssl", "version": "3.2.2", "release": "6.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libarchive": [{"name": "libarchive", "version": "3.5.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libevent": [{"name": "libevent", "version": "2.1.12", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_certmap": [{"name": "libsss_certmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz": [{"name": "xz", "version": "5.2.5", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "squashfs-tools": [{"name": "squashfs-tools", "version": "4.4", "release": "10.git1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcomps": [{"name": "libcomps", "version": "0.1.18", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libutempter": [{"name": "libutempter", "version": "1.2.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux-utils": [{"name": "libselinux-utils", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3-cli": [{"name": "libnl3-cli", "version": "3.9.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libteam": [{"name": "libteam", "version": "1.31", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "acl": [{"name": "acl", "version": "2.3.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-libs": [{"name": "gettext-libs", "version": "0.21", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext": [{"name": "gettext", "version": "0.21", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "attr": [{"name": "attr", "version": "2.5.1", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils": [{"name": "keyutils", "version": "1.6.3", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpfr": [{"name": "mpfr", "version": "4.1.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gawk": [{"name": "gawk", "version": "5.1.0", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpsl": [{"name": "libpsl", "version": "0.21.1", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libksba": [{"name": "libksba", "version": "1.5.1", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ethtool": [{"name": "ethtool", "version": "6.2", "release": "1.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "ipset-libs": [{"name": "ipset-libs", "version": "7.11", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ipset": [{"name": "ipset", "version": "7.11", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "groff-base": [{"name": "groff-base", "version": "1.22.4", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "snappy": [{"name": "snappy", "version": "1.1.8", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs-libs": [{"name": "e2fsprogs-libs", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libss": [{"name": "libss", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt-compat": [{"name": "libxcrypt-compat", "version": "4.4.18", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip-wheel": [{"name": "python3-pip-wheel", "version": "21.3.1", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python-unversioned-command": [{"name": "python-unversioned-command", "version": "3.9.19", "release": "8.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3": [{"name": "python3", "version": "3.9.19", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libs": [{"name": "python3-libs", "version": "3.9.19", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libcomps": [{"name": "python3-libcomps", "version": "0.1.18", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-six": [{"name": "python3-six", "version": "1.15.0", "release": "9.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dateutil": [{"name": "python3-dateutil", "version": "2.8.1", "release": "7.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "python3-systemd": [{"name": "python3-systemd", "version": "234", "release": "19.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng-python3": [{"name": "libcap-ng-python3", "version": "0.8.2", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pigz": [{"name": "pigz", "version": "2.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "hostname": [{"name": "hostname", "version": "3.23", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools-libs": [{"name": "kernel-tools-libs", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "less": [{"name": "less", "version": "590", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-rpm-macros": [{"name": "systemd-rpm-macros", "version": "252", "release": "47.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "c-ares": [{"name": "c-ares", "version": "1.19.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpio": [{"name": "cpio", "version": "2.13", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "diffutils": [{"name": "diffutils", "version": "3.7", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "inih": [{"name": "inih", "version": "49", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbrotli": [{"name": "libbrotli", "version": "1.0.9", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcbor": [{"name": "libcbor", "version": "0.7.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdaemon": [{"name": "libdaemon", "version": "0.14", "release": "23.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "teamd": [{"name": "teamd", "version": "1.31", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libeconf": [{"name": "libeconf", "version": "0.4.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpwquality": [{"name": "libpwquality", "version": "1.4.4", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pam": [{"name": "pam", "version": "1.5.1", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux": [{"name": "util-linux", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus": [{"name": "dbus", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "systemd-pam": [{"name": "systemd-pam", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd": [{"name": "systemd", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-common": [{"name": "dbus-common", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dbus-broker": [{"name": "dbus-broker", "version": "28", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-common": [{"name": "grub2-common", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "cronie-anacron": [{"name": "cronie-anacron", "version": "1.5.7", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cronie": [{"name": "cronie", "version": "1.5.7", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crontabs": [{"name": "crontabs", "version": "1.11", "release": "26.20190603git.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "openssh": [{"name": "openssh", "version": "8.7p1", "release": "43.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc-modules": [{"name": "grub2-pc-modules", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "authselect-libs": [{"name": "authselect-libs", "version": "1.2.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-libs": [{"name": "device-mapper-libs", "version": "1.02.198", "release": "2.el9", "epoch": 9, "arch": "x86_64", "source": "rpm"}], "device-mapper": [{"name": "device-mapper", "version": "1.02.198", "release": "2.el9", "epoch": 9, "arch": "x86_64", "source": "rpm"}], "grub2-tools-minimal": [{"name": "grub2-tools-minimal", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "cryptsetup-libs": [{"name": "cryptsetup-libs", "version": "2.7.2", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kpartx": [{"name": "kpartx", "version": "0.8.7", "release": "32.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-default-yama-scope": [{"name": "elfutils-default-yama-scope", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "elfutils-libs": [{"name": "elfutils-libs", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-service": [{"name": "initscripts-service", "version": "10.11.8", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iputils": [{"name": "iputils", "version": "20210202", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi": [{"name": "libkcapi", "version": "1.4.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hmaccalc": [{"name": "libkcapi-hmaccalc", "version": "1.4.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "logrotate": [{"name": "logrotate", "version": "3.18.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oddjob": [{"name": "oddjob", "version": "0.34.7", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oddjob-mkhomedir": [{"name": "oddjob-mkhomedir", "version": "0.34.7", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect": [{"name": "authselect", "version": "1.2.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kbd": [{"name": "kbd", "version": "2.4.0", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libndp": [{"name": "libndp", "version": "1.9", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfnetlink": [{"name": "libnfnetlink", "version": "1.0.1", "release": "21.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnetfilter_conntrack": [{"name": "libnetfilter_conntrack", "version": "1.0.9", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-libs": [{"name": "iptables-libs", "version": "1.8.10", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-nft": [{"name": "iptables-nft", "version": "1.8.10", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nftables": [{"name": "nftables", "version": "1.0.9", "release": "3.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-nftables": [{"name": "python3-nftables", "version": "1.0.9", "release": "3.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libnghttp2": [{"name": "libnghttp2", "version": "1.43.0", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpath_utils": [{"name": "libpath_utils", "version": "0.2.1", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libini_config": [{"name": "libini_config", "version": "1.3.1", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpipeline": [{"name": "libpipeline", "version": "1.5.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_nss_idmap": [{"name": "libsss_nss_idmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_sudo": [{"name": "libsss_sudo", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto": [{"name": "libverto", "version": "0.3.2", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "krb5-libs": [{"name": "krb5-libs", "version": "1.21.1", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-lib": [{"name": "cyrus-sasl-lib", "version": "2.1.27", "release": "21.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openldap": [{"name": "openldap", "version": "2.6.6", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libssh": [{"name": "libssh", "version": "0.10.4", "release": "13.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcurl": [{"name": "libcurl", "version": "7.76.1", "release": "31.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-debuginfod-client": [{"name": "elfutils-debuginfod-client", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils-gold": [{"name": "binutils-gold", "version": "2.35.2", "release": "54.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils": [{"name": "binutils", "version": "2.35.2", "release": "54.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tss": [{"name": "tpm2-tss", "version": "3.2.3", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-udev": [{"name": "systemd-udev", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut": [{"name": "dracut", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-modules-core": [{"name": "kernel-modules-core", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-core": [{"name": "kernel-core", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-modules": [{"name": "kernel-modules", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-squash": [{"name": "dracut-squash", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfido2": [{"name": "libfido2", "version": "1.13.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "os-prober": [{"name": "os-prober", "version": "1.77", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools": [{"name": "grub2-tools", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "ima-evm-utils": [{"name": "ima-evm-utils", "version": "1.5", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "curl": [{"name": "curl", "version": "7.76.1", "release": "31.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm": [{"name": "rpm", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-libs": [{"name": "rpm-libs", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grubby": [{"name": "grubby", "version": "8.40", "release": "63.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsolv": [{"name": "libsolv", "version": "0.7.24", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "policycoreutils": [{"name": "policycoreutils", "version": "3.6", "release": "2.1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "selinux-policy": [{"name": "selinux-policy", "version": "38.1.45", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "selinux-policy-targeted": [{"name": "selinux-policy-targeted", "version": "38.1.45", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpm-build-libs": [{"name": "rpm-build-libs", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-systemd-inhibit": [{"name": "rpm-plugin-systemd-inhibit", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-client": [{"name": "sssd-client", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libyaml": [{"name": "libyaml", "version": "0.2.5", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lmdb-libs": [{"name": "lmdb-libs", "version": "0.9.29", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libldb": [{"name": "libldb", "version": "2.9.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-common": [{"name": "sssd-common", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nettle": [{"name": "nettle", "version": "3.9.1", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnutls": [{"name": "gnutls", "version": "3.8.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glib2": [{"name": "glib2", "version": "2.68.4", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dbus": [{"name": "python3-dbus", "version": "1.2.18", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-libnm": [{"name": "NetworkManager-libnm", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "NetworkManager": [{"name": "NetworkManager", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libmodulemd": [{"name": "libmodulemd", "version": "2.13.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gobject-introspection": [{"name": "gobject-introspection", "version": "1.68.0", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gobject-base-noarch": [{"name": "python3-gobject-base-noarch", "version": "3.40.1", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-gobject-base": [{"name": "python3-gobject-base", "version": "3.40.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-firewall": [{"name": "python3-firewall", "version": "1.3.4", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libuser": [{"name": "libuser", "version": "0.63", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "npth": [{"name": "npth", "version": "1.6", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2": [{"name": "gnupg2", "version": "2.3.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gpgme": [{"name": "gpgme", "version": "1.15.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "librepo": [{"name": "librepo", "version": "1.14.5", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdnf": [{"name": "libdnf", "version": "0.69.0", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libdnf": [{"name": "python3-libdnf", "version": "0.69.0", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-hawkey": [{"name": "python3-hawkey", "version": "0.69.0", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gpg": [{"name": "python3-gpg", "version": "1.15.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sign-libs": [{"name": "rpm-sign-libs", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpm": [{"name": "python3-rpm", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dnf": [{"name": "python3-dnf", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf": [{"name": "dnf", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dnf-plugins-core": [{"name": "python3-dnf-plugins-core", "version": "4.3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "oniguruma": [{"name": "oniguruma", "version": "6.9.6", "release": "1.el9.6", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jq": [{"name": "jq", "version": "1.6", "release": "17.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-network": [{"name": "dracut-network", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pciutils-libs": [{"name": "pciutils-libs", "version": "3.7.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sg3_utils-libs": [{"name": "sg3_utils-libs", "version": "1.47", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "slang": [{"name": "slang", "version": "2.3.2", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "newt": [{"name": "newt", "version": "0.52.21", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "userspace-rcu": [{"name": "userspace-rcu", "version": "0.12.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libestr": [{"name": "libestr", "version": "0.1.11", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfastjson": [{"name": "libfastjson", "version": "0.99.9", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsyslog-logrotate": [{"name": "rsyslog-logrotate", "version": "8.2310.0", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsyslog": [{"name": "rsyslog", "version": "8.2310.0", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "liburing": [{"name": "liburing", "version": "2.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "langpacks-core-en": [{"name": "langpacks-core-en", "version": "3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-en": [{"name": "langpacks-en", "version": "3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "qemu-guest-agent": [{"name": "qemu-guest-agent", "version": "9.0.0", "release": "10.el9", "epoch": 17, "arch": "x86_64", "source": "rpm"}], "xfsprogs": [{"name": "xfsprogs", "version": "6.4.0", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-tui": [{"name": "NetworkManager-tui", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "sg3_utils": [{"name": "sg3_utils", "version": "1.47", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools": [{"name": "kernel-tools", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kexec-tools": [{"name": "kexec-tools", "version": "2.0.27", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dnf-plugins-core": [{"name": "dnf-plugins-core", "version": "4.3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "yum": [{"name": "yum", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "passwd": [{"name": "passwd", "version": "0.80", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "firewalld": [{"name": "firewalld", "version": "1.3.4", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "NetworkManager-team": [{"name": "NetworkManager-team", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "initscripts-rename-device": [{"name": "initscripts-rename-device", "version": "10.11.8", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "irqbalance": [{"name": "irqbalance", "version": "1.9.4", "release": "1.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "chrony": [{"name": "chrony", "version": "4.6", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-kcm": [{"name": "sssd-kcm", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-selinux": [{"name": "rpm-plugin-selinux", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies-scripts": [{"name": "crypto-policies-scripts", "version": "20240828", "release": "2.git626aa59.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpm-plugin-audit": [{"name": "rpm-plugin-audit", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc": [{"name": "grub2-pc", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "openssh-clients": [{"name": "openssh-clients", "version": "8.7p1", "release": "43.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel": [{"name": "kernel", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-config-rescue": [{"name": "dracut-config-rescue", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo": [{"name": "sudo", "version": "1.9.5p2", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit": [{"name": "audit", "version": "3.1.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-server": [{"name": "openssh-server", "version": "8.7p1", "release": "43.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "man-db": [{"name": "man-db", "version": "2.9.3", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute-tc": [{"name": "iproute-tc", "version": "6.2.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect-compat": [{"name": "authselect-compat", "version": "1.2.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "parted": [{"name": "parted", "version": "3.5", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "microcode_ctl": [{"name": "microcode_ctl", "version": "20240531", "release": "1.el9", "epoch": 4, "arch": "noarch", "source": "rpm"}], "python3-libselinux": [{"name": "python3-libselinux", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs": [{"name": "e2fsprogs", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "prefixdevname": [{"name": "prefixdevname", "version": "0.1.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-minimal": [{"name": "vim-minimal", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "lshw": [{"name": "lshw", "version": "B.02.19.2", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses": [{"name": "ncurses", "version": "6.2", "release": "10.20210508.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsysfs": [{"name": "libsysfs", "version": "2.1.1", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lsscsi": [{"name": "lsscsi", "version": "0.32", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iwl100-firmware": [{"name": "iwl100-firmware", "version": "39.31.5.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl1000-firmware": [{"name": "iwl1000-firmware", "version": "39.31.5.1", "release": "146.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "iwl105-firmware": [{"name": "iwl105-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl135-firmware": [{"name": "iwl135-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl2000-firmware": [{"name": "iwl2000-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl2030-firmware": [{"name": "iwl2030-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl3160-firmware": [{"name": "iwl3160-firmware", "version": "25.30.13.0", "release": "146.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "iwl5000-firmware": [{"name": "iwl5000-firmware", "version": "8.83.5.1_1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl5150-firmware": [{"name": "iwl5150-firmware", "version": "8.24.2.2", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl6000g2a-firmware": [{"name": "iwl6000g2a-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl6050-firmware": [{"name": "iwl6050-firmware", "version": "41.28.5.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl7260-firmware": [{"name": "iwl7260-firmware", "version": "25.30.13.0", "release": "146.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "rootfiles": [{"name": "rootfiles", "version": "8.1", "release": "31.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "gpg-pubkey": [{"name": "gpg-pubkey", "version": "3228467c", "release": "613798eb", "epoch": null, "arch": null, "source": "rpm"}, {"name": "gpg-pubkey", "version": "8483c65d", "release": "5ccc5b19", "epoch": null, "arch": null, "source": "rpm"}], "epel-release": [{"name": "epel-release", "version": "9", "release": "8.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "nspr": [{"name": "nspr", "version": "4.35.0", "release": "14.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "boost-system": [{"name": "boost-system", "version": "1.75.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-util": [{"name": "nss-util", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "make": [{"name": "make", "version": "4.3", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "m4": [{"name": "m4", "version": "1.4.19", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmpc": [{"name": "libmpc", "version": "1.2.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "unzip": [{"name": "unzip", "version": "6.0", "release": "57.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "avahi-libs": [{"name": "avahi-libs", "version": "0.8", "release": "21.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zip": [{"name": "zip", "version": "3.0", "release": "35.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpp": [{"name": "cpp", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bison": [{"name": "bison", "version": "3.7.4", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "flex": [{"name": "flex", "version": "2.6.4", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-softokn-freebl": [{"name": "nss-softokn-freebl", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-softokn": [{"name": "nss-softokn", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss": [{"name": "nss", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-sysinit": [{"name": "nss-sysinit", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "boost-filesystem": [{"name": "boost-filesystem", "version": "1.75.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "boost-thread": [{"name": "boost-thread", "version": "1.75.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Digest": [{"name": "perl-Digest", "version": "1.19", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Digest-MD5": [{"name": "perl-Digest-MD5", "version": "2.58", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-B": [{"name": "perl-B", "version": "1.80", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-FileHandle": [{"name": "perl-FileHandle", "version": "2.03", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Data-Dumper": [{"name": "perl-Data-Dumper", "version": "2.174", "release": "462.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-libnet": [{"name": "perl-libnet", "version": "3.13", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-base": [{"name": "perl-base", "version": "2.27", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-URI": [{"name": "perl-URI", "version": "5.09", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-AutoLoader": [{"name": "perl-AutoLoader", "version": "5.74", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Mozilla-CA": [{"name": "perl-Mozilla-CA", "version": "20200520", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-if": [{"name": "perl-if", "version": "0.60.800", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-IP": [{"name": "perl-IO-Socket-IP", "version": "0.41", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Time-Local": [{"name": "perl-Time-Local", "version": "1.300", "release": "7.el9", "epoch": 2, "arch": "noarch", "source": "rpm"}], "perl-File-Path": [{"name": "perl-File-Path", "version": "2.18", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Escapes": [{"name": "perl-Pod-Escapes", "version": "1.07", "release": "460.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Text-Tabs+Wrap": [{"name": "perl-Text-Tabs+Wrap", "version": "2013.0523", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-SSL": [{"name": "perl-IO-Socket-SSL", "version": "2.073", "release": "2.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Net-SSLeay": [{"name": "perl-Net-SSLeay", "version": "1.94", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Class-Struct": [{"name": "perl-Class-Struct", "version": "0.66", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-POSIX": [{"name": "perl-POSIX", "version": "1.94", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Term-ANSIColor": [{"name": "perl-Term-ANSIColor", "version": "5.01", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-IPC-Open3": [{"name": "perl-IPC-Open3", "version": "1.21", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-subs": [{"name": "perl-subs", "version": "1.03", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-Temp": [{"name": "perl-File-Temp", "version": "0.231.100", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Term-Cap": [{"name": "perl-Term-Cap", "version": "1.17", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Simple": [{"name": "perl-Pod-Simple", "version": "3.42", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-HTTP-Tiny": [{"name": "perl-HTTP-Tiny", "version": "0.076", "release": "462.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Socket": [{"name": "perl-Socket", "version": "2.031", "release": "4.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-SelectSaver": [{"name": "perl-SelectSaver", "version": "1.02", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Symbol": [{"name": "perl-Symbol", "version": "1.08", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-stat": [{"name": "perl-File-stat", "version": "1.09", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-podlators": [{"name": "perl-podlators", "version": "4.14", "release": "460.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Pod-Perldoc": [{"name": "perl-Pod-Perldoc", "version": "3.28.01", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Fcntl": [{"name": "perl-Fcntl", "version": "1.13", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Text-ParseWords": [{"name": "perl-Text-ParseWords", "version": "3.30", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-mro": [{"name": "perl-mro", "version": "1.23", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IO": [{"name": "perl-IO", "version": "1.43", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-overloading": [{"name": "perl-overloading", "version": "0.02", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Pod-Usage": [{"name": "perl-Pod-Usage", "version": "2.01", "release": "4.el9", "epoch": 4, "arch": "noarch", "source": "rpm"}], "perl-Errno": [{"name": "perl-Errno", "version": "1.30", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Basename": [{"name": "perl-File-Basename", "version": "2.85", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Std": [{"name": "perl-Getopt-Std", "version": "1.12", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-MIME-Base64": [{"name": "perl-MIME-Base64", "version": "3.16", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Scalar-List-Utils": [{"name": "perl-Scalar-List-Utils", "version": "1.56", "release": "462.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-constant": [{"name": "perl-constant", "version": "1.33", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Storable": [{"name": "perl-Storable", "version": "3.21", "release": "460.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "perl-overload": [{"name": "perl-overload", "version": "1.31", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-parent": [{"name": "perl-parent", "version": "0.238", "release": "460.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-vars": [{"name": "perl-vars", "version": "1.05", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Long": [{"name": "perl-Getopt-Long", "version": "2.52", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Carp": [{"name": "perl-Carp", "version": "1.50", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Exporter": [{"name": "perl-Exporter", "version": "5.74", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-NDBM_File": [{"name": "perl-NDBM_File", "version": "1.15", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-PathTools": [{"name": "perl-PathTools", "version": "3.78", "release": "461.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Encode": [{"name": "perl-Encode", "version": "3.08", "release": "462.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-libs": [{"name": "perl-libs", "version": "5.32.1", "release": "481.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-interpreter": [{"name": "perl-interpreter", "version": "5.32.1", "release": "481.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "aspell": [{"name": "aspell", "version": "0.60.8", "release": "8.el9", "epoch": 12, "arch": "x86_64", "source": "rpm"}], "tbb": [{"name": "tbb", "version": "2020.3", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dyninst": [{"name": "dyninst", "version": "12.1.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemtap-runtime": [{"name": "systemtap-runtime", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-headers": [{"name": "kernel-headers", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-headers": [{"name": "glibc-headers", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "strace": [{"name": "strace", "version": "5.18", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pkgconf-m4": [{"name": "pkgconf-m4", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libpkgconf": [{"name": "libpkgconf", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pkgconf": [{"name": "pkgconf", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pkgconf-pkg-config": [{"name": "pkgconf-pkg-config", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd-devel": [{"name": "libzstd-devel", "version": "1.5.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib-devel": [{"name": "zlib-devel", "version": "1.2.11", "release": "41.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf-devel": [{"name": "elfutils-libelf-devel", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-devel": [{"name": "glibc-devel", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt-devel": [{"name": "libxcrypt-devel", "version": "4.4.18", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gcc": [{"name": "gcc", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssl-devel": [{"name": "openssl-devel", "version": "3.2.2", "release": "6.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-devel": [{"name": "kernel-devel", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-devel": [{"name": "xz-devel", "version": "5.2.5", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-devel": [{"name": "elfutils-devel", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemtap-devel": [{"name": "systemtap-devel", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "efivar-libs": [{"name": "efivar-libs", "version": "38", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mokutil": [{"name": "mokutil", "version": "0.6.0", "release": "4.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "systemtap-client": [{"name": "systemtap-client", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemtap": [{"name": "systemtap", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "qa-tools": [{"name": "qa-tools", "version": "4.1", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libtirpc": [{"name": "libtirpc", "version": "1.3.3", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "git-core": [{"name": "git-core", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfsidmap": [{"name": "libnfsidmap", "version": "2.5.4", "release": "27.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "git-core-doc": [{"name": "git-core-doc", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpcbind": [{"name": "rpcbind", "version": "1.2.6", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wget": [{"name": "wget", "version": "1.21.1", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-lib": [{"name": "perl-lib", "version": "0.65", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Find": [{"name": "perl-File-Find", "version": "1.37", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Error": [{"name": "perl-Error", "version": "0.17029", "release": "7.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-DynaLoader": [{"name": "perl-DynaLoader", "version": "1.47", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-TermReadKey": [{"name": "perl-TermReadKey", "version": "2.38", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxslt": [{"name": "libxslt", "version": "1.1.34", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-lxml": [{"name": "python3-lxml", "version": "4.6.5", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gpm-libs": [{"name": "gpm-libs", "version": "1.20.7", "release": "29.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "emacs-filesystem": [{"name": "emacs-filesystem", "version": "27.2", "release": "10.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Git": [{"name": "perl-Git", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "git": [{"name": "git", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "yum-utils": [{"name": "yum-utils", "version": "4.3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "vim-filesystem": [{"name": "vim-filesystem", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "noarch", "source": "rpm"}], "vim-common": [{"name": "vim-common", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "time": [{"name": "time", "version": "1.9", "release": "18.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tar": [{"name": "tar", "version": "1.34", "release": "7.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "quota-nls": [{"name": "quota-nls", "version": "4.09", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "quota": [{"name": "quota", "version": "4.09", "release": "4.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-pyyaml": [{"name": "python3-pyyaml", "version": "5.4.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libev": [{"name": "libev", "version": "4.33", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto-libev": [{"name": "libverto-libev", "version": "0.3.2", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gssproxy": [{"name": "gssproxy", "version": "0.8.4", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nfs-utils": [{"name": "nfs-utils", "version": "2.5.4", "release": "27.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "bc": [{"name": "bc", "version": "1.07.1", "release": "14.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "beakerlib-redhat": [{"name": "beakerlib-redhat", "version": "1", "release": "35.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "beakerlib": [{"name": "beakerlib", "version": "1.29.3", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "restraint": [{"name": "restraint", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "restraint-rhts": [{"name": "restraint-rhts", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-enhanced": [{"name": "vim-enhanced", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "sssd-nfs-idmap": [{"name": "sssd-nfs-idmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsync": [{"name": "rsync", "version": "3.2.3", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools": [{"name": "python3-setuptools", "version": "53.0.0", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-distro": [{"name": "python3-distro", "version": "1.5.0", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-idna": [{"name": "python3-idna", "version": "2.10", "release": "7.el9.1", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-setools": [{"name": "python3-setools", "version": "4.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pytz": [{"name": "python3-pytz", "version": "2021.1", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-babel": [{"name": "python3-babel", "version": "2.9.1", "release": "2.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyserial": [{"name": "python3-pyserial", "version": "3.4", "release": "12.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyrsistent": [{"name": "python3-pyrsistent", "version": "0.17.3", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-prettytable": [{"name": "python3-prettytable", "version": "0.7.2", "release": "27.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-oauthlib": [{"name": "python3-oauthlib", "version": "3.1.1", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-netifaces": [{"name": "python3-netifaces", "version": "0.10.6", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-markupsafe": [{"name": "python3-markupsafe", "version": "1.1.1", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jinja2": [{"name": "python3-jinja2", "version": "2.11.3", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libsemanage": [{"name": "python3-libsemanage", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jsonpointer": [{"name": "python3-jsonpointer", "version": "2.0", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonpatch": [{"name": "python3-jsonpatch", "version": "1.21", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-configobj": [{"name": "python3-configobj", "version": "5.0.6", "release": "25.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-audit": [{"name": "python3-audit", "version": "3.1.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-attrs": [{"name": "python3-attrs", "version": "20.3.0", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema": [{"name": "python3-jsonschema", "version": "3.2.0", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libmaxminddb": [{"name": "libmaxminddb", "version": "1.5.2", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "geolite2-country": [{"name": "geolite2-country", "version": "20191217", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "geolite2-city": [{"name": "geolite2-city", "version": "20191217", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "ipcalc": [{"name": "ipcalc", "version": "1.0.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdisk": [{"name": "gdisk", "version": "1.0.7", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "checkpolicy": [{"name": "checkpolicy", "version": "3.6", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-policycoreutils": [{"name": "python3-policycoreutils", "version": "3.6", "release": "2.1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pysocks": [{"name": "python3-pysocks", "version": "1.7.1", "release": "12.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-urllib3": [{"name": "python3-urllib3", "version": "1.26.5", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-chardet": [{"name": "python3-chardet", "version": "4.0.0", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-requests": [{"name": "python3-requests", "version": "2.25.1", "release": "8.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "dhcp-common": [{"name": "dhcp-common", "version": "4.4.2", "release": "19.b1.el9", "epoch": 12, "arch": "noarch", "source": "rpm"}], "dhcp-client": [{"name": "dhcp-client", "version": "4.4.2", "release": "19.b1.el9", "epoch": 12, "arch": "x86_64", "source": "rpm"}], "cloud-init": [{"name": "cloud-init", "version": "23.4", "release": "19.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "cloud-utils-growpart": [{"name": "cloud-utils-growpart", "version": "0.33", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jitterentropy": [{"name": "jitterentropy", "version": "3.5.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rng-tools": [{"name": "rng-tools", "version": "6.16", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip": [{"name": "python3-pip", "version": "21.3.1", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "hostapd": [{"name": "hostapd", "version": "2.10", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wpa_supplicant": [{"name": "wpa_supplicant", "version": "2.10", "release": "5.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dnsmasq": [{"name": "dnsmasq", "version": "2.85", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}]}}, "invocation": {"module_args": {"manager": ["auto"], "strategy": "first"}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204077.92835: done with _execute_module (package_facts, {'_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'package_facts', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204076.8755333-14802-254494334316654/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204077.92871: _low_level_execute_command(): starting 13118 1727204077.92880: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204076.8755333-14802-254494334316654/ > /dev/null 2>&1 && sleep 0' 13118 1727204077.93614: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204077.93635: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204077.93652: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204077.93673: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204077.93727: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204077.93744: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204077.93760: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204077.93783: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204077.93801: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204077.93813: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204077.93826: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204077.93848: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204077.93866: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204077.93880: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204077.93891: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204077.93911: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204077.93995: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204077.94025: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204077.94047: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204077.94136: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204077.96085: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204077.96090: stdout chunk (state=3): >>><<< 13118 1727204077.96093: stderr chunk (state=3): >>><<< 13118 1727204077.96771: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204077.96775: handler run complete 13118 1727204077.97028: variable 'ansible_facts' from source: unknown 13118 1727204077.97518: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204077.99672: variable 'ansible_facts' from source: unknown 13118 1727204078.00154: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204078.00961: attempt loop complete, returning result 13118 1727204078.00983: _execute() done 13118 1727204078.00991: dumping result to json 13118 1727204078.01224: done dumping result, returning 13118 1727204078.01242: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check which packages are installed [0affcd87-79f5-56a3-0a64-00000000018e] 13118 1727204078.01254: sending task result for task 0affcd87-79f5-56a3-0a64-00000000018e 13118 1727204078.03648: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000018e 13118 1727204078.03651: WORKER PROCESS EXITING ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13118 1727204078.03734: no more pending results, returning what we have 13118 1727204078.03737: results queue empty 13118 1727204078.03738: checking for any_errors_fatal 13118 1727204078.03742: done checking for any_errors_fatal 13118 1727204078.03743: checking for max_fail_percentage 13118 1727204078.03744: done checking for max_fail_percentage 13118 1727204078.03745: checking to see if all hosts have failed and the running result is not ok 13118 1727204078.03746: done checking to see if all hosts have failed 13118 1727204078.03746: getting the remaining hosts for this loop 13118 1727204078.03748: done getting the remaining hosts for this loop 13118 1727204078.03751: getting the next task for host managed-node2 13118 1727204078.03758: done getting next task for host managed-node2 13118 1727204078.03762: ^ task is: TASK: fedora.linux_system_roles.network : Print network provider 13118 1727204078.03766: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204078.03776: getting variables 13118 1727204078.03777: in VariableManager get_vars() 13118 1727204078.03809: Calling all_inventory to load vars for managed-node2 13118 1727204078.03812: Calling groups_inventory to load vars for managed-node2 13118 1727204078.03814: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204078.03823: Calling all_plugins_play to load vars for managed-node2 13118 1727204078.03826: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204078.03829: Calling groups_plugins_play to load vars for managed-node2 13118 1727204078.06135: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204078.08254: done with get_vars() 13118 1727204078.08290: done getting variables 13118 1727204078.08356: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Print network provider] ************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:7 Tuesday 24 September 2024 14:54:38 -0400 (0:00:01.265) 0:00:15.383 ***** 13118 1727204078.08392: entering _queue_task() for managed-node2/debug 13118 1727204078.09988: worker is 1 (out of 1 available) 13118 1727204078.10001: exiting _queue_task() for managed-node2/debug 13118 1727204078.10015: done queuing things up, now waiting for results queue to drain 13118 1727204078.10017: waiting for pending results... 13118 1727204078.10550: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Print network provider 13118 1727204078.10689: in run() - task 0affcd87-79f5-56a3-0a64-000000000027 13118 1727204078.10712: variable 'ansible_search_path' from source: unknown 13118 1727204078.10720: variable 'ansible_search_path' from source: unknown 13118 1727204078.10769: calling self._execute() 13118 1727204078.10857: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204078.10872: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204078.10891: variable 'omit' from source: magic vars 13118 1727204078.11286: variable 'ansible_distribution_major_version' from source: facts 13118 1727204078.11303: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204078.11314: variable 'omit' from source: magic vars 13118 1727204078.11380: variable 'omit' from source: magic vars 13118 1727204078.11486: variable 'network_provider' from source: set_fact 13118 1727204078.11507: variable 'omit' from source: magic vars 13118 1727204078.11559: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204078.11602: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204078.11632: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204078.11660: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204078.11680: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204078.11712: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204078.11719: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204078.11726: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204078.11839: Set connection var ansible_timeout to 10 13118 1727204078.11857: Set connection var ansible_pipelining to False 13118 1727204078.11867: Set connection var ansible_connection to ssh 13118 1727204078.11882: Set connection var ansible_shell_executable to /bin/sh 13118 1727204078.11891: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204078.11897: Set connection var ansible_shell_type to sh 13118 1727204078.11921: variable 'ansible_shell_executable' from source: unknown 13118 1727204078.11931: variable 'ansible_connection' from source: unknown 13118 1727204078.11938: variable 'ansible_module_compression' from source: unknown 13118 1727204078.11945: variable 'ansible_shell_type' from source: unknown 13118 1727204078.11950: variable 'ansible_shell_executable' from source: unknown 13118 1727204078.11956: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204078.11962: variable 'ansible_pipelining' from source: unknown 13118 1727204078.11971: variable 'ansible_timeout' from source: unknown 13118 1727204078.11980: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204078.12134: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204078.12151: variable 'omit' from source: magic vars 13118 1727204078.12159: starting attempt loop 13118 1727204078.12168: running the handler 13118 1727204078.12219: handler run complete 13118 1727204078.12242: attempt loop complete, returning result 13118 1727204078.12250: _execute() done 13118 1727204078.12257: dumping result to json 13118 1727204078.12266: done dumping result, returning 13118 1727204078.12278: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Print network provider [0affcd87-79f5-56a3-0a64-000000000027] 13118 1727204078.12289: sending task result for task 0affcd87-79f5-56a3-0a64-000000000027 13118 1727204078.12399: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000027 13118 1727204078.12406: WORKER PROCESS EXITING ok: [managed-node2] => {} MSG: Using network provider: nm 13118 1727204078.12473: no more pending results, returning what we have 13118 1727204078.12477: results queue empty 13118 1727204078.12477: checking for any_errors_fatal 13118 1727204078.12488: done checking for any_errors_fatal 13118 1727204078.12488: checking for max_fail_percentage 13118 1727204078.12490: done checking for max_fail_percentage 13118 1727204078.12490: checking to see if all hosts have failed and the running result is not ok 13118 1727204078.12491: done checking to see if all hosts have failed 13118 1727204078.12492: getting the remaining hosts for this loop 13118 1727204078.12493: done getting the remaining hosts for this loop 13118 1727204078.12497: getting the next task for host managed-node2 13118 1727204078.12504: done getting next task for host managed-node2 13118 1727204078.12507: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider 13118 1727204078.12511: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204078.12520: getting variables 13118 1727204078.12527: in VariableManager get_vars() 13118 1727204078.12570: Calling all_inventory to load vars for managed-node2 13118 1727204078.12573: Calling groups_inventory to load vars for managed-node2 13118 1727204078.12575: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204078.12585: Calling all_plugins_play to load vars for managed-node2 13118 1727204078.12587: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204078.12589: Calling groups_plugins_play to load vars for managed-node2 13118 1727204078.14686: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204078.18295: done with get_vars() 13118 1727204078.18331: done getting variables 13118 1727204078.18648: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:11 Tuesday 24 September 2024 14:54:38 -0400 (0:00:00.104) 0:00:15.488 ***** 13118 1727204078.18804: entering _queue_task() for managed-node2/fail 13118 1727204078.18806: Creating lock for fail 13118 1727204078.19601: worker is 1 (out of 1 available) 13118 1727204078.19615: exiting _queue_task() for managed-node2/fail 13118 1727204078.19628: done queuing things up, now waiting for results queue to drain 13118 1727204078.19630: waiting for pending results... 13118 1727204078.21038: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider 13118 1727204078.21184: in run() - task 0affcd87-79f5-56a3-0a64-000000000028 13118 1727204078.21206: variable 'ansible_search_path' from source: unknown 13118 1727204078.21215: variable 'ansible_search_path' from source: unknown 13118 1727204078.21341: calling self._execute() 13118 1727204078.21547: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204078.21559: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204078.21575: variable 'omit' from source: magic vars 13118 1727204078.22396: variable 'ansible_distribution_major_version' from source: facts 13118 1727204078.22419: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204078.22674: variable 'network_state' from source: role '' defaults 13118 1727204078.22689: Evaluated conditional (network_state != {}): False 13118 1727204078.22696: when evaluation is False, skipping this task 13118 1727204078.22704: _execute() done 13118 1727204078.22719: dumping result to json 13118 1727204078.22728: done dumping result, returning 13118 1727204078.22742: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider [0affcd87-79f5-56a3-0a64-000000000028] 13118 1727204078.22752: sending task result for task 0affcd87-79f5-56a3-0a64-000000000028 skipping: [managed-node2] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13118 1727204078.22975: no more pending results, returning what we have 13118 1727204078.22979: results queue empty 13118 1727204078.22980: checking for any_errors_fatal 13118 1727204078.22986: done checking for any_errors_fatal 13118 1727204078.22986: checking for max_fail_percentage 13118 1727204078.22988: done checking for max_fail_percentage 13118 1727204078.22989: checking to see if all hosts have failed and the running result is not ok 13118 1727204078.22990: done checking to see if all hosts have failed 13118 1727204078.22990: getting the remaining hosts for this loop 13118 1727204078.22991: done getting the remaining hosts for this loop 13118 1727204078.22995: getting the next task for host managed-node2 13118 1727204078.23002: done getting next task for host managed-node2 13118 1727204078.23005: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 13118 1727204078.23008: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204078.23024: getting variables 13118 1727204078.23025: in VariableManager get_vars() 13118 1727204078.23072: Calling all_inventory to load vars for managed-node2 13118 1727204078.23076: Calling groups_inventory to load vars for managed-node2 13118 1727204078.23078: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204078.23091: Calling all_plugins_play to load vars for managed-node2 13118 1727204078.23095: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204078.23098: Calling groups_plugins_play to load vars for managed-node2 13118 1727204078.23813: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000028 13118 1727204078.23817: WORKER PROCESS EXITING 13118 1727204078.25989: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204078.29159: done with get_vars() 13118 1727204078.29202: done getting variables 13118 1727204078.29278: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:18 Tuesday 24 September 2024 14:54:38 -0400 (0:00:00.105) 0:00:15.593 ***** 13118 1727204078.29322: entering _queue_task() for managed-node2/fail 13118 1727204078.29675: worker is 1 (out of 1 available) 13118 1727204078.29689: exiting _queue_task() for managed-node2/fail 13118 1727204078.29701: done queuing things up, now waiting for results queue to drain 13118 1727204078.29703: waiting for pending results... 13118 1727204078.30005: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 13118 1727204078.30495: in run() - task 0affcd87-79f5-56a3-0a64-000000000029 13118 1727204078.30515: variable 'ansible_search_path' from source: unknown 13118 1727204078.30523: variable 'ansible_search_path' from source: unknown 13118 1727204078.30575: calling self._execute() 13118 1727204078.30663: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204078.30682: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204078.30697: variable 'omit' from source: magic vars 13118 1727204078.31085: variable 'ansible_distribution_major_version' from source: facts 13118 1727204078.31102: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204078.31228: variable 'network_state' from source: role '' defaults 13118 1727204078.31245: Evaluated conditional (network_state != {}): False 13118 1727204078.31251: when evaluation is False, skipping this task 13118 1727204078.31258: _execute() done 13118 1727204078.31266: dumping result to json 13118 1727204078.31274: done dumping result, returning 13118 1727204078.31285: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 [0affcd87-79f5-56a3-0a64-000000000029] 13118 1727204078.31298: sending task result for task 0affcd87-79f5-56a3-0a64-000000000029 skipping: [managed-node2] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13118 1727204078.31487: no more pending results, returning what we have 13118 1727204078.31492: results queue empty 13118 1727204078.31493: checking for any_errors_fatal 13118 1727204078.31500: done checking for any_errors_fatal 13118 1727204078.31501: checking for max_fail_percentage 13118 1727204078.31503: done checking for max_fail_percentage 13118 1727204078.31503: checking to see if all hosts have failed and the running result is not ok 13118 1727204078.31504: done checking to see if all hosts have failed 13118 1727204078.31505: getting the remaining hosts for this loop 13118 1727204078.31506: done getting the remaining hosts for this loop 13118 1727204078.31511: getting the next task for host managed-node2 13118 1727204078.31519: done getting next task for host managed-node2 13118 1727204078.31524: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later 13118 1727204078.31528: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204078.31552: getting variables 13118 1727204078.31554: in VariableManager get_vars() 13118 1727204078.31601: Calling all_inventory to load vars for managed-node2 13118 1727204078.31604: Calling groups_inventory to load vars for managed-node2 13118 1727204078.31606: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204078.31620: Calling all_plugins_play to load vars for managed-node2 13118 1727204078.31623: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204078.31626: Calling groups_plugins_play to load vars for managed-node2 13118 1727204078.32603: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000029 13118 1727204078.32607: WORKER PROCESS EXITING 13118 1727204078.33579: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204078.36949: done with get_vars() 13118 1727204078.36978: done getting variables 13118 1727204078.37160: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:25 Tuesday 24 September 2024 14:54:38 -0400 (0:00:00.078) 0:00:15.671 ***** 13118 1727204078.37198: entering _queue_task() for managed-node2/fail 13118 1727204078.37860: worker is 1 (out of 1 available) 13118 1727204078.37988: exiting _queue_task() for managed-node2/fail 13118 1727204078.38002: done queuing things up, now waiting for results queue to drain 13118 1727204078.38003: waiting for pending results... 13118 1727204078.38818: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later 13118 1727204078.39110: in run() - task 0affcd87-79f5-56a3-0a64-00000000002a 13118 1727204078.39202: variable 'ansible_search_path' from source: unknown 13118 1727204078.39210: variable 'ansible_search_path' from source: unknown 13118 1727204078.39253: calling self._execute() 13118 1727204078.39476: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204078.39487: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204078.39501: variable 'omit' from source: magic vars 13118 1727204078.40338: variable 'ansible_distribution_major_version' from source: facts 13118 1727204078.40357: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204078.40678: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204078.46314: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204078.46448: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204078.46638: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204078.46681: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204078.46837: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204078.46944: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.47063: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.47099: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.47185: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.47269: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.47493: variable 'ansible_distribution_major_version' from source: facts 13118 1727204078.47515: Evaluated conditional (ansible_distribution_major_version | int > 9): False 13118 1727204078.47578: when evaluation is False, skipping this task 13118 1727204078.47589: _execute() done 13118 1727204078.47597: dumping result to json 13118 1727204078.47604: done dumping result, returning 13118 1727204078.47616: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later [0affcd87-79f5-56a3-0a64-00000000002a] 13118 1727204078.47626: sending task result for task 0affcd87-79f5-56a3-0a64-00000000002a skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_distribution_major_version | int > 9", "skip_reason": "Conditional result was False" } 13118 1727204078.47785: no more pending results, returning what we have 13118 1727204078.47789: results queue empty 13118 1727204078.47790: checking for any_errors_fatal 13118 1727204078.47796: done checking for any_errors_fatal 13118 1727204078.47797: checking for max_fail_percentage 13118 1727204078.47799: done checking for max_fail_percentage 13118 1727204078.47800: checking to see if all hosts have failed and the running result is not ok 13118 1727204078.47801: done checking to see if all hosts have failed 13118 1727204078.47801: getting the remaining hosts for this loop 13118 1727204078.47803: done getting the remaining hosts for this loop 13118 1727204078.47807: getting the next task for host managed-node2 13118 1727204078.47815: done getting next task for host managed-node2 13118 1727204078.47819: ^ task is: TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces 13118 1727204078.47822: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204078.47840: getting variables 13118 1727204078.47843: in VariableManager get_vars() 13118 1727204078.47890: Calling all_inventory to load vars for managed-node2 13118 1727204078.47893: Calling groups_inventory to load vars for managed-node2 13118 1727204078.47896: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204078.47907: Calling all_plugins_play to load vars for managed-node2 13118 1727204078.47909: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204078.47913: Calling groups_plugins_play to load vars for managed-node2 13118 1727204078.49394: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000002a 13118 1727204078.49398: WORKER PROCESS EXITING 13118 1727204078.51079: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204078.54722: done with get_vars() 13118 1727204078.54760: done getting variables 13118 1727204078.55040: Loading ActionModule 'dnf' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/dnf.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:36 Tuesday 24 September 2024 14:54:38 -0400 (0:00:00.178) 0:00:15.850 ***** 13118 1727204078.55079: entering _queue_task() for managed-node2/dnf 13118 1727204078.55758: worker is 1 (out of 1 available) 13118 1727204078.55913: exiting _queue_task() for managed-node2/dnf 13118 1727204078.55925: done queuing things up, now waiting for results queue to drain 13118 1727204078.55927: waiting for pending results... 13118 1727204078.56679: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces 13118 1727204078.57090: in run() - task 0affcd87-79f5-56a3-0a64-00000000002b 13118 1727204078.57109: variable 'ansible_search_path' from source: unknown 13118 1727204078.57113: variable 'ansible_search_path' from source: unknown 13118 1727204078.57148: calling self._execute() 13118 1727204078.57342: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204078.57348: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204078.57358: variable 'omit' from source: magic vars 13118 1727204078.58959: variable 'ansible_distribution_major_version' from source: facts 13118 1727204078.59677: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204078.59882: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204078.64357: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204078.64433: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204078.64482: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204078.64507: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204078.64535: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204078.64609: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.64636: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.64661: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.64703: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.64720: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.64835: variable 'ansible_distribution' from source: facts 13118 1727204078.64838: variable 'ansible_distribution_major_version' from source: facts 13118 1727204078.64856: Evaluated conditional (ansible_distribution == 'Fedora' or ansible_distribution_major_version | int > 7): True 13118 1727204078.64971: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204078.65088: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.65108: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.65132: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.65165: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.65177: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.65213: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.65234: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.65254: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.65290: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.65303: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.65341: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.65366: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.65391: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.65434: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.65445: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.65596: variable 'network_connections' from source: task vars 13118 1727204078.65609: variable 'controller_profile' from source: play vars 13118 1727204078.65674: variable 'controller_profile' from source: play vars 13118 1727204078.65682: variable 'controller_device' from source: play vars 13118 1727204078.65740: variable 'controller_device' from source: play vars 13118 1727204078.65751: variable 'port1_profile' from source: play vars 13118 1727204078.65835: variable 'port1_profile' from source: play vars 13118 1727204078.65845: variable 'dhcp_interface1' from source: play vars 13118 1727204078.65901: variable 'dhcp_interface1' from source: play vars 13118 1727204078.65907: variable 'controller_profile' from source: play vars 13118 1727204078.65993: variable 'controller_profile' from source: play vars 13118 1727204078.66001: variable 'port2_profile' from source: play vars 13118 1727204078.66063: variable 'port2_profile' from source: play vars 13118 1727204078.66135: variable 'dhcp_interface2' from source: play vars 13118 1727204078.66180: variable 'dhcp_interface2' from source: play vars 13118 1727204078.66183: variable 'controller_profile' from source: play vars 13118 1727204078.66243: variable 'controller_profile' from source: play vars 13118 1727204078.66561: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204078.66566: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204078.66569: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204078.66572: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204078.66863: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204078.67076: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204078.67079: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204078.67082: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.67084: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204078.67087: variable '__network_team_connections_defined' from source: role '' defaults 13118 1727204078.67089: variable 'network_connections' from source: task vars 13118 1727204078.67091: variable 'controller_profile' from source: play vars 13118 1727204078.67094: variable 'controller_profile' from source: play vars 13118 1727204078.67095: variable 'controller_device' from source: play vars 13118 1727204078.67135: variable 'controller_device' from source: play vars 13118 1727204078.67144: variable 'port1_profile' from source: play vars 13118 1727204078.67334: variable 'port1_profile' from source: play vars 13118 1727204078.67337: variable 'dhcp_interface1' from source: play vars 13118 1727204078.67340: variable 'dhcp_interface1' from source: play vars 13118 1727204078.67342: variable 'controller_profile' from source: play vars 13118 1727204078.67351: variable 'controller_profile' from source: play vars 13118 1727204078.67371: variable 'port2_profile' from source: play vars 13118 1727204078.67433: variable 'port2_profile' from source: play vars 13118 1727204078.67437: variable 'dhcp_interface2' from source: play vars 13118 1727204078.67494: variable 'dhcp_interface2' from source: play vars 13118 1727204078.67500: variable 'controller_profile' from source: play vars 13118 1727204078.67556: variable 'controller_profile' from source: play vars 13118 1727204078.67599: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13118 1727204078.67603: when evaluation is False, skipping this task 13118 1727204078.67605: _execute() done 13118 1727204078.67608: dumping result to json 13118 1727204078.67610: done dumping result, returning 13118 1727204078.67613: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces [0affcd87-79f5-56a3-0a64-00000000002b] 13118 1727204078.67625: sending task result for task 0affcd87-79f5-56a3-0a64-00000000002b 13118 1727204078.67917: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000002b 13118 1727204078.67920: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13118 1727204078.67967: no more pending results, returning what we have 13118 1727204078.67970: results queue empty 13118 1727204078.67972: checking for any_errors_fatal 13118 1727204078.67978: done checking for any_errors_fatal 13118 1727204078.67979: checking for max_fail_percentage 13118 1727204078.67981: done checking for max_fail_percentage 13118 1727204078.67982: checking to see if all hosts have failed and the running result is not ok 13118 1727204078.67983: done checking to see if all hosts have failed 13118 1727204078.67983: getting the remaining hosts for this loop 13118 1727204078.67985: done getting the remaining hosts for this loop 13118 1727204078.67989: getting the next task for host managed-node2 13118 1727204078.67995: done getting next task for host managed-node2 13118 1727204078.67999: ^ task is: TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces 13118 1727204078.68002: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204078.68017: getting variables 13118 1727204078.68019: in VariableManager get_vars() 13118 1727204078.68063: Calling all_inventory to load vars for managed-node2 13118 1727204078.68068: Calling groups_inventory to load vars for managed-node2 13118 1727204078.68070: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204078.68081: Calling all_plugins_play to load vars for managed-node2 13118 1727204078.68084: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204078.68087: Calling groups_plugins_play to load vars for managed-node2 13118 1727204078.70015: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204078.72299: done with get_vars() 13118 1727204078.72334: done getting variables redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf 13118 1727204078.72416: Loading ActionModule 'ansible_collections.ansible.builtin.plugins.action.dnf' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/dnf.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:48 Tuesday 24 September 2024 14:54:38 -0400 (0:00:00.173) 0:00:16.024 ***** 13118 1727204078.72451: entering _queue_task() for managed-node2/yum 13118 1727204078.72453: Creating lock for yum 13118 1727204078.72784: worker is 1 (out of 1 available) 13118 1727204078.72813: exiting _queue_task() for managed-node2/yum 13118 1727204078.72827: done queuing things up, now waiting for results queue to drain 13118 1727204078.72828: waiting for pending results... 13118 1727204078.73838: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces 13118 1727204078.73950: in run() - task 0affcd87-79f5-56a3-0a64-00000000002c 13118 1727204078.73963: variable 'ansible_search_path' from source: unknown 13118 1727204078.73968: variable 'ansible_search_path' from source: unknown 13118 1727204078.74004: calling self._execute() 13118 1727204078.74084: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204078.74090: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204078.74100: variable 'omit' from source: magic vars 13118 1727204078.74585: variable 'ansible_distribution_major_version' from source: facts 13118 1727204078.74605: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204078.74785: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204078.77438: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204078.77521: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204078.77558: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204078.77598: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204078.77629: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204078.77710: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.77742: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.77771: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.77812: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.77831: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.77931: variable 'ansible_distribution_major_version' from source: facts 13118 1727204078.77953: Evaluated conditional (ansible_distribution_major_version | int < 8): False 13118 1727204078.77960: when evaluation is False, skipping this task 13118 1727204078.77970: _execute() done 13118 1727204078.77977: dumping result to json 13118 1727204078.77984: done dumping result, returning 13118 1727204078.77995: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces [0affcd87-79f5-56a3-0a64-00000000002c] 13118 1727204078.78005: sending task result for task 0affcd87-79f5-56a3-0a64-00000000002c skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_distribution_major_version | int < 8", "skip_reason": "Conditional result was False" } 13118 1727204078.78171: no more pending results, returning what we have 13118 1727204078.78175: results queue empty 13118 1727204078.78176: checking for any_errors_fatal 13118 1727204078.78181: done checking for any_errors_fatal 13118 1727204078.78182: checking for max_fail_percentage 13118 1727204078.78183: done checking for max_fail_percentage 13118 1727204078.78184: checking to see if all hosts have failed and the running result is not ok 13118 1727204078.78185: done checking to see if all hosts have failed 13118 1727204078.78186: getting the remaining hosts for this loop 13118 1727204078.78187: done getting the remaining hosts for this loop 13118 1727204078.78190: getting the next task for host managed-node2 13118 1727204078.78197: done getting next task for host managed-node2 13118 1727204078.78201: ^ task is: TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces 13118 1727204078.78203: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204078.78222: getting variables 13118 1727204078.78223: in VariableManager get_vars() 13118 1727204078.78266: Calling all_inventory to load vars for managed-node2 13118 1727204078.78269: Calling groups_inventory to load vars for managed-node2 13118 1727204078.78271: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204078.78282: Calling all_plugins_play to load vars for managed-node2 13118 1727204078.78285: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204078.78288: Calling groups_plugins_play to load vars for managed-node2 13118 1727204078.78809: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000002c 13118 1727204078.78812: WORKER PROCESS EXITING 13118 1727204078.79676: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204078.81313: done with get_vars() 13118 1727204078.81337: done getting variables 13118 1727204078.81401: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:60 Tuesday 24 September 2024 14:54:38 -0400 (0:00:00.089) 0:00:16.114 ***** 13118 1727204078.81440: entering _queue_task() for managed-node2/fail 13118 1727204078.81788: worker is 1 (out of 1 available) 13118 1727204078.81802: exiting _queue_task() for managed-node2/fail 13118 1727204078.81815: done queuing things up, now waiting for results queue to drain 13118 1727204078.81816: waiting for pending results... 13118 1727204078.82156: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces 13118 1727204078.82301: in run() - task 0affcd87-79f5-56a3-0a64-00000000002d 13118 1727204078.82322: variable 'ansible_search_path' from source: unknown 13118 1727204078.82330: variable 'ansible_search_path' from source: unknown 13118 1727204078.82376: calling self._execute() 13118 1727204078.82465: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204078.82488: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204078.82526: variable 'omit' from source: magic vars 13118 1727204078.82975: variable 'ansible_distribution_major_version' from source: facts 13118 1727204078.82992: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204078.83117: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204078.83313: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204078.86034: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204078.86113: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204078.86155: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204078.86199: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204078.86228: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204078.86313: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.86346: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.86382: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.86433: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.86454: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.86506: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.86538: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.86569: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.86613: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.86637: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.86684: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.86712: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.86746: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.86791: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.86812: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.86997: variable 'network_connections' from source: task vars 13118 1727204078.87014: variable 'controller_profile' from source: play vars 13118 1727204078.87095: variable 'controller_profile' from source: play vars 13118 1727204078.87108: variable 'controller_device' from source: play vars 13118 1727204078.87177: variable 'controller_device' from source: play vars 13118 1727204078.87193: variable 'port1_profile' from source: play vars 13118 1727204078.87255: variable 'port1_profile' from source: play vars 13118 1727204078.87274: variable 'dhcp_interface1' from source: play vars 13118 1727204078.87335: variable 'dhcp_interface1' from source: play vars 13118 1727204078.87347: variable 'controller_profile' from source: play vars 13118 1727204078.87414: variable 'controller_profile' from source: play vars 13118 1727204078.87427: variable 'port2_profile' from source: play vars 13118 1727204078.87493: variable 'port2_profile' from source: play vars 13118 1727204078.87505: variable 'dhcp_interface2' from source: play vars 13118 1727204078.87569: variable 'dhcp_interface2' from source: play vars 13118 1727204078.87580: variable 'controller_profile' from source: play vars 13118 1727204078.87645: variable 'controller_profile' from source: play vars 13118 1727204078.87724: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204078.87922: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204078.87963: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204078.88002: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204078.88043: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204078.88092: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204078.88117: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204078.88149: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.88181: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204078.88257: variable '__network_team_connections_defined' from source: role '' defaults 13118 1727204078.88506: variable 'network_connections' from source: task vars 13118 1727204078.88515: variable 'controller_profile' from source: play vars 13118 1727204078.88582: variable 'controller_profile' from source: play vars 13118 1727204078.88593: variable 'controller_device' from source: play vars 13118 1727204078.88655: variable 'controller_device' from source: play vars 13118 1727204078.88670: variable 'port1_profile' from source: play vars 13118 1727204078.88735: variable 'port1_profile' from source: play vars 13118 1727204078.88746: variable 'dhcp_interface1' from source: play vars 13118 1727204078.88815: variable 'dhcp_interface1' from source: play vars 13118 1727204078.88825: variable 'controller_profile' from source: play vars 13118 1727204078.88889: variable 'controller_profile' from source: play vars 13118 1727204078.88901: variable 'port2_profile' from source: play vars 13118 1727204078.88966: variable 'port2_profile' from source: play vars 13118 1727204078.88978: variable 'dhcp_interface2' from source: play vars 13118 1727204078.89044: variable 'dhcp_interface2' from source: play vars 13118 1727204078.89055: variable 'controller_profile' from source: play vars 13118 1727204078.89120: variable 'controller_profile' from source: play vars 13118 1727204078.89158: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13118 1727204078.89169: when evaluation is False, skipping this task 13118 1727204078.89177: _execute() done 13118 1727204078.89183: dumping result to json 13118 1727204078.89190: done dumping result, returning 13118 1727204078.89200: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces [0affcd87-79f5-56a3-0a64-00000000002d] 13118 1727204078.89210: sending task result for task 0affcd87-79f5-56a3-0a64-00000000002d 13118 1727204078.89319: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000002d 13118 1727204078.89327: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13118 1727204078.89388: no more pending results, returning what we have 13118 1727204078.89392: results queue empty 13118 1727204078.89393: checking for any_errors_fatal 13118 1727204078.89401: done checking for any_errors_fatal 13118 1727204078.89401: checking for max_fail_percentage 13118 1727204078.89403: done checking for max_fail_percentage 13118 1727204078.89404: checking to see if all hosts have failed and the running result is not ok 13118 1727204078.89405: done checking to see if all hosts have failed 13118 1727204078.89406: getting the remaining hosts for this loop 13118 1727204078.89407: done getting the remaining hosts for this loop 13118 1727204078.89411: getting the next task for host managed-node2 13118 1727204078.89419: done getting next task for host managed-node2 13118 1727204078.89423: ^ task is: TASK: fedora.linux_system_roles.network : Install packages 13118 1727204078.89425: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204078.89440: getting variables 13118 1727204078.89442: in VariableManager get_vars() 13118 1727204078.89487: Calling all_inventory to load vars for managed-node2 13118 1727204078.89490: Calling groups_inventory to load vars for managed-node2 13118 1727204078.89492: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204078.89503: Calling all_plugins_play to load vars for managed-node2 13118 1727204078.89506: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204078.89509: Calling groups_plugins_play to load vars for managed-node2 13118 1727204078.91213: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204078.93153: done with get_vars() 13118 1727204078.93187: done getting variables 13118 1727204078.93251: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install packages] ******************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:73 Tuesday 24 September 2024 14:54:38 -0400 (0:00:00.118) 0:00:16.232 ***** 13118 1727204078.93290: entering _queue_task() for managed-node2/package 13118 1727204078.93597: worker is 1 (out of 1 available) 13118 1727204078.93609: exiting _queue_task() for managed-node2/package 13118 1727204078.93623: done queuing things up, now waiting for results queue to drain 13118 1727204078.93624: waiting for pending results... 13118 1727204078.93901: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Install packages 13118 1727204078.94037: in run() - task 0affcd87-79f5-56a3-0a64-00000000002e 13118 1727204078.94055: variable 'ansible_search_path' from source: unknown 13118 1727204078.94063: variable 'ansible_search_path' from source: unknown 13118 1727204078.94104: calling self._execute() 13118 1727204078.94191: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204078.94201: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204078.94214: variable 'omit' from source: magic vars 13118 1727204078.94582: variable 'ansible_distribution_major_version' from source: facts 13118 1727204078.94599: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204078.94808: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204078.95082: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204078.95130: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204078.95173: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204078.95211: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204078.95329: variable 'network_packages' from source: role '' defaults 13118 1727204078.95442: variable '__network_provider_setup' from source: role '' defaults 13118 1727204078.95456: variable '__network_service_name_default_nm' from source: role '' defaults 13118 1727204078.95529: variable '__network_service_name_default_nm' from source: role '' defaults 13118 1727204078.95543: variable '__network_packages_default_nm' from source: role '' defaults 13118 1727204078.95614: variable '__network_packages_default_nm' from source: role '' defaults 13118 1727204078.95844: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204078.97977: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204078.98061: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204078.98111: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204078.98150: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204078.98188: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204078.98273: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.98313: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.98345: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.98420: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.98447: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.98498: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.98531: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.98574: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.98635: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.98657: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.98984: variable '__network_packages_default_gobject_packages' from source: role '' defaults 13118 1727204078.99137: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.99168: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.99203: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.99260: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.99305: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.99404: variable 'ansible_python' from source: facts 13118 1727204078.99435: variable '__network_packages_default_wpa_supplicant' from source: role '' defaults 13118 1727204078.99533: variable '__network_wpa_supplicant_required' from source: role '' defaults 13118 1727204078.99627: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13118 1727204078.99758: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.99788: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204078.99818: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204078.99868: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204078.99887: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204078.99937: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204078.99979: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204079.00009: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204079.00056: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204079.00082: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204079.00342: variable 'network_connections' from source: task vars 13118 1727204079.00354: variable 'controller_profile' from source: play vars 13118 1727204079.00488: variable 'controller_profile' from source: play vars 13118 1727204079.00498: variable 'controller_device' from source: play vars 13118 1727204079.00608: variable 'controller_device' from source: play vars 13118 1727204079.00617: variable 'port1_profile' from source: play vars 13118 1727204079.00698: variable 'port1_profile' from source: play vars 13118 1727204079.00707: variable 'dhcp_interface1' from source: play vars 13118 1727204079.00781: variable 'dhcp_interface1' from source: play vars 13118 1727204079.00786: variable 'controller_profile' from source: play vars 13118 1727204079.00856: variable 'controller_profile' from source: play vars 13118 1727204079.00864: variable 'port2_profile' from source: play vars 13118 1727204079.00935: variable 'port2_profile' from source: play vars 13118 1727204079.00944: variable 'dhcp_interface2' from source: play vars 13118 1727204079.01012: variable 'dhcp_interface2' from source: play vars 13118 1727204079.01019: variable 'controller_profile' from source: play vars 13118 1727204079.01089: variable 'controller_profile' from source: play vars 13118 1727204079.01148: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204079.01169: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204079.01190: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204079.01212: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204079.01256: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204079.01433: variable 'network_connections' from source: task vars 13118 1727204079.01439: variable 'controller_profile' from source: play vars 13118 1727204079.01510: variable 'controller_profile' from source: play vars 13118 1727204079.01517: variable 'controller_device' from source: play vars 13118 1727204079.01590: variable 'controller_device' from source: play vars 13118 1727204079.01599: variable 'port1_profile' from source: play vars 13118 1727204079.01669: variable 'port1_profile' from source: play vars 13118 1727204079.01677: variable 'dhcp_interface1' from source: play vars 13118 1727204079.01746: variable 'dhcp_interface1' from source: play vars 13118 1727204079.01753: variable 'controller_profile' from source: play vars 13118 1727204079.01824: variable 'controller_profile' from source: play vars 13118 1727204079.01833: variable 'port2_profile' from source: play vars 13118 1727204079.01905: variable 'port2_profile' from source: play vars 13118 1727204079.01910: variable 'dhcp_interface2' from source: play vars 13118 1727204079.01981: variable 'dhcp_interface2' from source: play vars 13118 1727204079.01989: variable 'controller_profile' from source: play vars 13118 1727204079.02058: variable 'controller_profile' from source: play vars 13118 1727204079.02100: variable '__network_packages_default_wireless' from source: role '' defaults 13118 1727204079.02170: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204079.02576: variable 'network_connections' from source: task vars 13118 1727204079.02587: variable 'controller_profile' from source: play vars 13118 1727204079.02665: variable 'controller_profile' from source: play vars 13118 1727204079.02681: variable 'controller_device' from source: play vars 13118 1727204079.02744: variable 'controller_device' from source: play vars 13118 1727204079.02767: variable 'port1_profile' from source: play vars 13118 1727204079.02833: variable 'port1_profile' from source: play vars 13118 1727204079.02845: variable 'dhcp_interface1' from source: play vars 13118 1727204079.02919: variable 'dhcp_interface1' from source: play vars 13118 1727204079.02930: variable 'controller_profile' from source: play vars 13118 1727204079.03004: variable 'controller_profile' from source: play vars 13118 1727204079.03016: variable 'port2_profile' from source: play vars 13118 1727204079.03113: variable 'port2_profile' from source: play vars 13118 1727204079.03125: variable 'dhcp_interface2' from source: play vars 13118 1727204079.03263: variable 'dhcp_interface2' from source: play vars 13118 1727204079.03276: variable 'controller_profile' from source: play vars 13118 1727204079.03387: variable 'controller_profile' from source: play vars 13118 1727204079.03424: variable '__network_packages_default_team' from source: role '' defaults 13118 1727204079.03491: variable '__network_team_connections_defined' from source: role '' defaults 13118 1727204079.03699: variable 'network_connections' from source: task vars 13118 1727204079.03702: variable 'controller_profile' from source: play vars 13118 1727204079.03752: variable 'controller_profile' from source: play vars 13118 1727204079.03759: variable 'controller_device' from source: play vars 13118 1727204079.03805: variable 'controller_device' from source: play vars 13118 1727204079.03812: variable 'port1_profile' from source: play vars 13118 1727204079.03860: variable 'port1_profile' from source: play vars 13118 1727204079.03868: variable 'dhcp_interface1' from source: play vars 13118 1727204079.03912: variable 'dhcp_interface1' from source: play vars 13118 1727204079.03917: variable 'controller_profile' from source: play vars 13118 1727204079.03968: variable 'controller_profile' from source: play vars 13118 1727204079.03971: variable 'port2_profile' from source: play vars 13118 1727204079.04018: variable 'port2_profile' from source: play vars 13118 1727204079.04023: variable 'dhcp_interface2' from source: play vars 13118 1727204079.04072: variable 'dhcp_interface2' from source: play vars 13118 1727204079.04076: variable 'controller_profile' from source: play vars 13118 1727204079.04122: variable 'controller_profile' from source: play vars 13118 1727204079.04171: variable '__network_service_name_default_initscripts' from source: role '' defaults 13118 1727204079.04214: variable '__network_service_name_default_initscripts' from source: role '' defaults 13118 1727204079.04220: variable '__network_packages_default_initscripts' from source: role '' defaults 13118 1727204079.04265: variable '__network_packages_default_initscripts' from source: role '' defaults 13118 1727204079.04403: variable '__network_packages_default_initscripts_bridge' from source: role '' defaults 13118 1727204079.04701: variable 'network_connections' from source: task vars 13118 1727204079.04704: variable 'controller_profile' from source: play vars 13118 1727204079.04753: variable 'controller_profile' from source: play vars 13118 1727204079.04761: variable 'controller_device' from source: play vars 13118 1727204079.04803: variable 'controller_device' from source: play vars 13118 1727204079.04810: variable 'port1_profile' from source: play vars 13118 1727204079.04858: variable 'port1_profile' from source: play vars 13118 1727204079.04863: variable 'dhcp_interface1' from source: play vars 13118 1727204079.04905: variable 'dhcp_interface1' from source: play vars 13118 1727204079.04910: variable 'controller_profile' from source: play vars 13118 1727204079.04954: variable 'controller_profile' from source: play vars 13118 1727204079.04961: variable 'port2_profile' from source: play vars 13118 1727204079.05003: variable 'port2_profile' from source: play vars 13118 1727204079.05008: variable 'dhcp_interface2' from source: play vars 13118 1727204079.05052: variable 'dhcp_interface2' from source: play vars 13118 1727204079.05057: variable 'controller_profile' from source: play vars 13118 1727204079.05101: variable 'controller_profile' from source: play vars 13118 1727204079.05107: variable 'ansible_distribution' from source: facts 13118 1727204079.05110: variable '__network_rh_distros' from source: role '' defaults 13118 1727204079.05116: variable 'ansible_distribution_major_version' from source: facts 13118 1727204079.05138: variable '__network_packages_default_initscripts_network_scripts' from source: role '' defaults 13118 1727204079.05251: variable 'ansible_distribution' from source: facts 13118 1727204079.05270: variable '__network_rh_distros' from source: role '' defaults 13118 1727204079.05274: variable 'ansible_distribution_major_version' from source: facts 13118 1727204079.05294: variable '__network_packages_default_initscripts_dhcp_client' from source: role '' defaults 13118 1727204079.05450: variable 'ansible_distribution' from source: facts 13118 1727204079.05453: variable '__network_rh_distros' from source: role '' defaults 13118 1727204079.05458: variable 'ansible_distribution_major_version' from source: facts 13118 1727204079.05510: variable 'network_provider' from source: set_fact 13118 1727204079.05545: variable 'ansible_facts' from source: unknown 13118 1727204079.06396: Evaluated conditional (not network_packages is subset(ansible_facts.packages.keys())): False 13118 1727204079.06405: when evaluation is False, skipping this task 13118 1727204079.06414: _execute() done 13118 1727204079.06422: dumping result to json 13118 1727204079.06429: done dumping result, returning 13118 1727204079.06441: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Install packages [0affcd87-79f5-56a3-0a64-00000000002e] 13118 1727204079.06452: sending task result for task 0affcd87-79f5-56a3-0a64-00000000002e 13118 1727204079.06584: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000002e skipping: [managed-node2] => { "changed": false, "false_condition": "not network_packages is subset(ansible_facts.packages.keys())", "skip_reason": "Conditional result was False" } 13118 1727204079.06638: no more pending results, returning what we have 13118 1727204079.06642: results queue empty 13118 1727204079.06642: checking for any_errors_fatal 13118 1727204079.06648: done checking for any_errors_fatal 13118 1727204079.06649: checking for max_fail_percentage 13118 1727204079.06651: done checking for max_fail_percentage 13118 1727204079.06652: checking to see if all hosts have failed and the running result is not ok 13118 1727204079.06653: done checking to see if all hosts have failed 13118 1727204079.06654: getting the remaining hosts for this loop 13118 1727204079.06655: done getting the remaining hosts for this loop 13118 1727204079.06659: getting the next task for host managed-node2 13118 1727204079.06668: done getting next task for host managed-node2 13118 1727204079.06672: ^ task is: TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable 13118 1727204079.06675: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204079.06691: getting variables 13118 1727204079.06693: in VariableManager get_vars() 13118 1727204079.06736: Calling all_inventory to load vars for managed-node2 13118 1727204079.06739: Calling groups_inventory to load vars for managed-node2 13118 1727204079.06742: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204079.06753: Calling all_plugins_play to load vars for managed-node2 13118 1727204079.06755: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204079.06758: Calling groups_plugins_play to load vars for managed-node2 13118 1727204079.07708: WORKER PROCESS EXITING 13118 1727204079.07998: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204079.08919: done with get_vars() 13118 1727204079.08941: done getting variables 13118 1727204079.08989: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:85 Tuesday 24 September 2024 14:54:39 -0400 (0:00:00.157) 0:00:16.390 ***** 13118 1727204079.09015: entering _queue_task() for managed-node2/package 13118 1727204079.09307: worker is 1 (out of 1 available) 13118 1727204079.09342: exiting _queue_task() for managed-node2/package 13118 1727204079.09355: done queuing things up, now waiting for results queue to drain 13118 1727204079.09356: waiting for pending results... 13118 1727204079.10190: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable 13118 1727204079.10196: in run() - task 0affcd87-79f5-56a3-0a64-00000000002f 13118 1727204079.10199: variable 'ansible_search_path' from source: unknown 13118 1727204079.10202: variable 'ansible_search_path' from source: unknown 13118 1727204079.10204: calling self._execute() 13118 1727204079.10206: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204079.10209: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204079.10211: variable 'omit' from source: magic vars 13118 1727204079.10325: variable 'ansible_distribution_major_version' from source: facts 13118 1727204079.10331: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204079.10390: variable 'network_state' from source: role '' defaults 13118 1727204079.10400: Evaluated conditional (network_state != {}): False 13118 1727204079.10403: when evaluation is False, skipping this task 13118 1727204079.10406: _execute() done 13118 1727204079.10409: dumping result to json 13118 1727204079.10412: done dumping result, returning 13118 1727204079.10439: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable [0affcd87-79f5-56a3-0a64-00000000002f] 13118 1727204079.10443: sending task result for task 0affcd87-79f5-56a3-0a64-00000000002f 13118 1727204079.10536: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000002f 13118 1727204079.10539: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13118 1727204079.10637: no more pending results, returning what we have 13118 1727204079.10642: results queue empty 13118 1727204079.10643: checking for any_errors_fatal 13118 1727204079.10653: done checking for any_errors_fatal 13118 1727204079.10877: checking for max_fail_percentage 13118 1727204079.10880: done checking for max_fail_percentage 13118 1727204079.10881: checking to see if all hosts have failed and the running result is not ok 13118 1727204079.10882: done checking to see if all hosts have failed 13118 1727204079.10882: getting the remaining hosts for this loop 13118 1727204079.10884: done getting the remaining hosts for this loop 13118 1727204079.10888: getting the next task for host managed-node2 13118 1727204079.10910: done getting next task for host managed-node2 13118 1727204079.10915: ^ task is: TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable 13118 1727204079.10918: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=14, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204079.10933: getting variables 13118 1727204079.10935: in VariableManager get_vars() 13118 1727204079.10983: Calling all_inventory to load vars for managed-node2 13118 1727204079.10986: Calling groups_inventory to load vars for managed-node2 13118 1727204079.10989: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204079.10999: Calling all_plugins_play to load vars for managed-node2 13118 1727204079.11001: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204079.11004: Calling groups_plugins_play to load vars for managed-node2 13118 1727204079.12204: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204079.13123: done with get_vars() 13118 1727204079.13145: done getting variables 13118 1727204079.13192: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:96 Tuesday 24 September 2024 14:54:39 -0400 (0:00:00.042) 0:00:16.432 ***** 13118 1727204079.13219: entering _queue_task() for managed-node2/package 13118 1727204079.13487: worker is 1 (out of 1 available) 13118 1727204079.13498: exiting _queue_task() for managed-node2/package 13118 1727204079.13510: done queuing things up, now waiting for results queue to drain 13118 1727204079.13511: waiting for pending results... 13118 1727204079.13806: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable 13118 1727204079.13937: in run() - task 0affcd87-79f5-56a3-0a64-000000000030 13118 1727204079.13949: variable 'ansible_search_path' from source: unknown 13118 1727204079.13958: variable 'ansible_search_path' from source: unknown 13118 1727204079.13996: calling self._execute() 13118 1727204079.14083: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204079.14088: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204079.14100: variable 'omit' from source: magic vars 13118 1727204079.14480: variable 'ansible_distribution_major_version' from source: facts 13118 1727204079.14501: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204079.14624: variable 'network_state' from source: role '' defaults 13118 1727204079.14639: Evaluated conditional (network_state != {}): False 13118 1727204079.14643: when evaluation is False, skipping this task 13118 1727204079.14646: _execute() done 13118 1727204079.14648: dumping result to json 13118 1727204079.14651: done dumping result, returning 13118 1727204079.14654: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable [0affcd87-79f5-56a3-0a64-000000000030] 13118 1727204079.14670: sending task result for task 0affcd87-79f5-56a3-0a64-000000000030 13118 1727204079.14778: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000030 13118 1727204079.14781: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13118 1727204079.14833: no more pending results, returning what we have 13118 1727204079.14838: results queue empty 13118 1727204079.14839: checking for any_errors_fatal 13118 1727204079.14843: done checking for any_errors_fatal 13118 1727204079.14844: checking for max_fail_percentage 13118 1727204079.14846: done checking for max_fail_percentage 13118 1727204079.14847: checking to see if all hosts have failed and the running result is not ok 13118 1727204079.14848: done checking to see if all hosts have failed 13118 1727204079.14849: getting the remaining hosts for this loop 13118 1727204079.14850: done getting the remaining hosts for this loop 13118 1727204079.14854: getting the next task for host managed-node2 13118 1727204079.14862: done getting next task for host managed-node2 13118 1727204079.14868: ^ task is: TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces 13118 1727204079.14872: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204079.14889: getting variables 13118 1727204079.14891: in VariableManager get_vars() 13118 1727204079.14940: Calling all_inventory to load vars for managed-node2 13118 1727204079.14943: Calling groups_inventory to load vars for managed-node2 13118 1727204079.14946: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204079.14965: Calling all_plugins_play to load vars for managed-node2 13118 1727204079.14969: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204079.14972: Calling groups_plugins_play to load vars for managed-node2 13118 1727204079.15868: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204079.19679: done with get_vars() 13118 1727204079.19699: done getting variables 13118 1727204079.19768: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:109 Tuesday 24 September 2024 14:54:39 -0400 (0:00:00.065) 0:00:16.497 ***** 13118 1727204079.19789: entering _queue_task() for managed-node2/service 13118 1727204079.19791: Creating lock for service 13118 1727204079.20025: worker is 1 (out of 1 available) 13118 1727204079.20043: exiting _queue_task() for managed-node2/service 13118 1727204079.20056: done queuing things up, now waiting for results queue to drain 13118 1727204079.20058: waiting for pending results... 13118 1727204079.20238: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces 13118 1727204079.20335: in run() - task 0affcd87-79f5-56a3-0a64-000000000031 13118 1727204079.20347: variable 'ansible_search_path' from source: unknown 13118 1727204079.20351: variable 'ansible_search_path' from source: unknown 13118 1727204079.20387: calling self._execute() 13118 1727204079.20452: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204079.20457: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204079.20466: variable 'omit' from source: magic vars 13118 1727204079.20743: variable 'ansible_distribution_major_version' from source: facts 13118 1727204079.20752: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204079.20840: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204079.20975: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204079.22588: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204079.22642: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204079.22672: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204079.22698: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204079.22718: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204079.22779: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204079.22802: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204079.22821: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204079.22852: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204079.22864: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204079.22897: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204079.22914: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204079.22933: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204079.22959: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204079.22971: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204079.23000: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204079.23017: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204079.23035: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204079.23063: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204079.23073: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204079.23191: variable 'network_connections' from source: task vars 13118 1727204079.23201: variable 'controller_profile' from source: play vars 13118 1727204079.23256: variable 'controller_profile' from source: play vars 13118 1727204079.23265: variable 'controller_device' from source: play vars 13118 1727204079.23311: variable 'controller_device' from source: play vars 13118 1727204079.23320: variable 'port1_profile' from source: play vars 13118 1727204079.23366: variable 'port1_profile' from source: play vars 13118 1727204079.23372: variable 'dhcp_interface1' from source: play vars 13118 1727204079.23417: variable 'dhcp_interface1' from source: play vars 13118 1727204079.23422: variable 'controller_profile' from source: play vars 13118 1727204079.23467: variable 'controller_profile' from source: play vars 13118 1727204079.23473: variable 'port2_profile' from source: play vars 13118 1727204079.23517: variable 'port2_profile' from source: play vars 13118 1727204079.23522: variable 'dhcp_interface2' from source: play vars 13118 1727204079.23568: variable 'dhcp_interface2' from source: play vars 13118 1727204079.23575: variable 'controller_profile' from source: play vars 13118 1727204079.23619: variable 'controller_profile' from source: play vars 13118 1727204079.23669: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204079.23794: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204079.23824: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204079.23846: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204079.23868: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204079.23901: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204079.23916: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204079.23934: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204079.23954: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204079.24008: variable '__network_team_connections_defined' from source: role '' defaults 13118 1727204079.24161: variable 'network_connections' from source: task vars 13118 1727204079.24166: variable 'controller_profile' from source: play vars 13118 1727204079.24209: variable 'controller_profile' from source: play vars 13118 1727204079.24215: variable 'controller_device' from source: play vars 13118 1727204079.24259: variable 'controller_device' from source: play vars 13118 1727204079.24268: variable 'port1_profile' from source: play vars 13118 1727204079.24310: variable 'port1_profile' from source: play vars 13118 1727204079.24314: variable 'dhcp_interface1' from source: play vars 13118 1727204079.24356: variable 'dhcp_interface1' from source: play vars 13118 1727204079.24366: variable 'controller_profile' from source: play vars 13118 1727204079.24408: variable 'controller_profile' from source: play vars 13118 1727204079.24413: variable 'port2_profile' from source: play vars 13118 1727204079.24455: variable 'port2_profile' from source: play vars 13118 1727204079.24461: variable 'dhcp_interface2' from source: play vars 13118 1727204079.24506: variable 'dhcp_interface2' from source: play vars 13118 1727204079.24511: variable 'controller_profile' from source: play vars 13118 1727204079.24554: variable 'controller_profile' from source: play vars 13118 1727204079.24583: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13118 1727204079.24591: when evaluation is False, skipping this task 13118 1727204079.24594: _execute() done 13118 1727204079.24596: dumping result to json 13118 1727204079.24598: done dumping result, returning 13118 1727204079.24601: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces [0affcd87-79f5-56a3-0a64-000000000031] 13118 1727204079.24607: sending task result for task 0affcd87-79f5-56a3-0a64-000000000031 13118 1727204079.24705: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000031 13118 1727204079.24708: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13118 1727204079.24754: no more pending results, returning what we have 13118 1727204079.24758: results queue empty 13118 1727204079.24758: checking for any_errors_fatal 13118 1727204079.24766: done checking for any_errors_fatal 13118 1727204079.24767: checking for max_fail_percentage 13118 1727204079.24769: done checking for max_fail_percentage 13118 1727204079.24769: checking to see if all hosts have failed and the running result is not ok 13118 1727204079.24770: done checking to see if all hosts have failed 13118 1727204079.24771: getting the remaining hosts for this loop 13118 1727204079.24772: done getting the remaining hosts for this loop 13118 1727204079.24776: getting the next task for host managed-node2 13118 1727204079.24782: done getting next task for host managed-node2 13118 1727204079.24786: ^ task is: TASK: fedora.linux_system_roles.network : Enable and start NetworkManager 13118 1727204079.24788: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204079.24811: getting variables 13118 1727204079.24813: in VariableManager get_vars() 13118 1727204079.24855: Calling all_inventory to load vars for managed-node2 13118 1727204079.24858: Calling groups_inventory to load vars for managed-node2 13118 1727204079.24859: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204079.24870: Calling all_plugins_play to load vars for managed-node2 13118 1727204079.24873: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204079.24876: Calling groups_plugins_play to load vars for managed-node2 13118 1727204079.25688: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204079.26622: done with get_vars() 13118 1727204079.26644: done getting variables 13118 1727204079.26691: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable and start NetworkManager] ***** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:122 Tuesday 24 September 2024 14:54:39 -0400 (0:00:00.069) 0:00:16.567 ***** 13118 1727204079.26715: entering _queue_task() for managed-node2/service 13118 1727204079.26955: worker is 1 (out of 1 available) 13118 1727204079.26969: exiting _queue_task() for managed-node2/service 13118 1727204079.26984: done queuing things up, now waiting for results queue to drain 13118 1727204079.26985: waiting for pending results... 13118 1727204079.27171: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Enable and start NetworkManager 13118 1727204079.27259: in run() - task 0affcd87-79f5-56a3-0a64-000000000032 13118 1727204079.27272: variable 'ansible_search_path' from source: unknown 13118 1727204079.27275: variable 'ansible_search_path' from source: unknown 13118 1727204079.27305: calling self._execute() 13118 1727204079.27375: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204079.27379: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204079.27388: variable 'omit' from source: magic vars 13118 1727204079.27658: variable 'ansible_distribution_major_version' from source: facts 13118 1727204079.27670: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204079.27781: variable 'network_provider' from source: set_fact 13118 1727204079.27786: variable 'network_state' from source: role '' defaults 13118 1727204079.27795: Evaluated conditional (network_provider == "nm" or network_state != {}): True 13118 1727204079.27800: variable 'omit' from source: magic vars 13118 1727204079.27840: variable 'omit' from source: magic vars 13118 1727204079.27863: variable 'network_service_name' from source: role '' defaults 13118 1727204079.27912: variable 'network_service_name' from source: role '' defaults 13118 1727204079.27988: variable '__network_provider_setup' from source: role '' defaults 13118 1727204079.27991: variable '__network_service_name_default_nm' from source: role '' defaults 13118 1727204079.28038: variable '__network_service_name_default_nm' from source: role '' defaults 13118 1727204079.28050: variable '__network_packages_default_nm' from source: role '' defaults 13118 1727204079.28117: variable '__network_packages_default_nm' from source: role '' defaults 13118 1727204079.28271: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204079.30332: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204079.30430: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204079.30464: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204079.30496: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204079.30519: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204079.30585: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204079.30611: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204079.30632: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204079.30661: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204079.30676: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204079.30715: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204079.30733: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204079.30751: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204079.30779: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204079.30790: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204079.30949: variable '__network_packages_default_gobject_packages' from source: role '' defaults 13118 1727204079.31031: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204079.31052: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204079.31072: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204079.31097: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204079.31107: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204079.31177: variable 'ansible_python' from source: facts 13118 1727204079.31194: variable '__network_packages_default_wpa_supplicant' from source: role '' defaults 13118 1727204079.31259: variable '__network_wpa_supplicant_required' from source: role '' defaults 13118 1727204079.31317: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13118 1727204079.31406: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204079.31422: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204079.31442: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204079.31474: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204079.31483: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204079.31517: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204079.31538: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204079.31559: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204079.31586: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204079.31597: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204079.31695: variable 'network_connections' from source: task vars 13118 1727204079.31703: variable 'controller_profile' from source: play vars 13118 1727204079.31757: variable 'controller_profile' from source: play vars 13118 1727204079.31768: variable 'controller_device' from source: play vars 13118 1727204079.31822: variable 'controller_device' from source: play vars 13118 1727204079.31835: variable 'port1_profile' from source: play vars 13118 1727204079.31890: variable 'port1_profile' from source: play vars 13118 1727204079.31900: variable 'dhcp_interface1' from source: play vars 13118 1727204079.31951: variable 'dhcp_interface1' from source: play vars 13118 1727204079.31959: variable 'controller_profile' from source: play vars 13118 1727204079.32014: variable 'controller_profile' from source: play vars 13118 1727204079.32023: variable 'port2_profile' from source: play vars 13118 1727204079.32076: variable 'port2_profile' from source: play vars 13118 1727204079.32085: variable 'dhcp_interface2' from source: play vars 13118 1727204079.32141: variable 'dhcp_interface2' from source: play vars 13118 1727204079.32149: variable 'controller_profile' from source: play vars 13118 1727204079.32229: variable 'controller_profile' from source: play vars 13118 1727204079.32338: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204079.33086: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204079.33089: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204079.33092: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204079.33094: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204079.33097: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204079.33099: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204079.33194: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204079.33197: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204079.33200: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204079.33451: variable 'network_connections' from source: task vars 13118 1727204079.33456: variable 'controller_profile' from source: play vars 13118 1727204079.33529: variable 'controller_profile' from source: play vars 13118 1727204079.33543: variable 'controller_device' from source: play vars 13118 1727204079.33613: variable 'controller_device' from source: play vars 13118 1727204079.33628: variable 'port1_profile' from source: play vars 13118 1727204079.33699: variable 'port1_profile' from source: play vars 13118 1727204079.33709: variable 'dhcp_interface1' from source: play vars 13118 1727204079.33782: variable 'dhcp_interface1' from source: play vars 13118 1727204079.33792: variable 'controller_profile' from source: play vars 13118 1727204079.33863: variable 'controller_profile' from source: play vars 13118 1727204079.33875: variable 'port2_profile' from source: play vars 13118 1727204079.33947: variable 'port2_profile' from source: play vars 13118 1727204079.33955: variable 'dhcp_interface2' from source: play vars 13118 1727204079.34027: variable 'dhcp_interface2' from source: play vars 13118 1727204079.34040: variable 'controller_profile' from source: play vars 13118 1727204079.34110: variable 'controller_profile' from source: play vars 13118 1727204079.34160: variable '__network_packages_default_wireless' from source: role '' defaults 13118 1727204079.34240: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204079.34527: variable 'network_connections' from source: task vars 13118 1727204079.34531: variable 'controller_profile' from source: play vars 13118 1727204079.34605: variable 'controller_profile' from source: play vars 13118 1727204079.34612: variable 'controller_device' from source: play vars 13118 1727204079.34681: variable 'controller_device' from source: play vars 13118 1727204079.34689: variable 'port1_profile' from source: play vars 13118 1727204079.34759: variable 'port1_profile' from source: play vars 13118 1727204079.34767: variable 'dhcp_interface1' from source: play vars 13118 1727204079.34837: variable 'dhcp_interface1' from source: play vars 13118 1727204079.34843: variable 'controller_profile' from source: play vars 13118 1727204079.34912: variable 'controller_profile' from source: play vars 13118 1727204079.34918: variable 'port2_profile' from source: play vars 13118 1727204079.34989: variable 'port2_profile' from source: play vars 13118 1727204079.34995: variable 'dhcp_interface2' from source: play vars 13118 1727204079.35065: variable 'dhcp_interface2' from source: play vars 13118 1727204079.35072: variable 'controller_profile' from source: play vars 13118 1727204079.35141: variable 'controller_profile' from source: play vars 13118 1727204079.35167: variable '__network_packages_default_team' from source: role '' defaults 13118 1727204079.35247: variable '__network_team_connections_defined' from source: role '' defaults 13118 1727204079.35536: variable 'network_connections' from source: task vars 13118 1727204079.35539: variable 'controller_profile' from source: play vars 13118 1727204079.35609: variable 'controller_profile' from source: play vars 13118 1727204079.35616: variable 'controller_device' from source: play vars 13118 1727204079.35687: variable 'controller_device' from source: play vars 13118 1727204079.35698: variable 'port1_profile' from source: play vars 13118 1727204079.35766: variable 'port1_profile' from source: play vars 13118 1727204079.35772: variable 'dhcp_interface1' from source: play vars 13118 1727204079.35841: variable 'dhcp_interface1' from source: play vars 13118 1727204079.35847: variable 'controller_profile' from source: play vars 13118 1727204079.35932: variable 'controller_profile' from source: play vars 13118 1727204079.35942: variable 'port2_profile' from source: play vars 13118 1727204079.36010: variable 'port2_profile' from source: play vars 13118 1727204079.36018: variable 'dhcp_interface2' from source: play vars 13118 1727204079.36088: variable 'dhcp_interface2' from source: play vars 13118 1727204079.36098: variable 'controller_profile' from source: play vars 13118 1727204079.36171: variable 'controller_profile' from source: play vars 13118 1727204079.36242: variable '__network_service_name_default_initscripts' from source: role '' defaults 13118 1727204079.36312: variable '__network_service_name_default_initscripts' from source: role '' defaults 13118 1727204079.36323: variable '__network_packages_default_initscripts' from source: role '' defaults 13118 1727204079.36390: variable '__network_packages_default_initscripts' from source: role '' defaults 13118 1727204079.36609: variable '__network_packages_default_initscripts_bridge' from source: role '' defaults 13118 1727204079.37068: variable 'network_connections' from source: task vars 13118 1727204079.37078: variable 'controller_profile' from source: play vars 13118 1727204079.37139: variable 'controller_profile' from source: play vars 13118 1727204079.37153: variable 'controller_device' from source: play vars 13118 1727204079.37220: variable 'controller_device' from source: play vars 13118 1727204079.37240: variable 'port1_profile' from source: play vars 13118 1727204079.37304: variable 'port1_profile' from source: play vars 13118 1727204079.37317: variable 'dhcp_interface1' from source: play vars 13118 1727204079.37388: variable 'dhcp_interface1' from source: play vars 13118 1727204079.37400: variable 'controller_profile' from source: play vars 13118 1727204079.37469: variable 'controller_profile' from source: play vars 13118 1727204079.37483: variable 'port2_profile' from source: play vars 13118 1727204079.37547: variable 'port2_profile' from source: play vars 13118 1727204079.37563: variable 'dhcp_interface2' from source: play vars 13118 1727204079.37625: variable 'dhcp_interface2' from source: play vars 13118 1727204079.37640: variable 'controller_profile' from source: play vars 13118 1727204079.37706: variable 'controller_profile' from source: play vars 13118 1727204079.37720: variable 'ansible_distribution' from source: facts 13118 1727204079.37732: variable '__network_rh_distros' from source: role '' defaults 13118 1727204079.37743: variable 'ansible_distribution_major_version' from source: facts 13118 1727204079.37779: variable '__network_packages_default_initscripts_network_scripts' from source: role '' defaults 13118 1727204079.37968: variable 'ansible_distribution' from source: facts 13118 1727204079.37978: variable '__network_rh_distros' from source: role '' defaults 13118 1727204079.37992: variable 'ansible_distribution_major_version' from source: facts 13118 1727204079.38010: variable '__network_packages_default_initscripts_dhcp_client' from source: role '' defaults 13118 1727204079.38192: variable 'ansible_distribution' from source: facts 13118 1727204079.38206: variable '__network_rh_distros' from source: role '' defaults 13118 1727204079.38217: variable 'ansible_distribution_major_version' from source: facts 13118 1727204079.38261: variable 'network_provider' from source: set_fact 13118 1727204079.38291: variable 'omit' from source: magic vars 13118 1727204079.38333: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204079.38369: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204079.38394: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204079.38421: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204079.38440: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204079.38476: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204079.38484: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204079.38492: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204079.38603: Set connection var ansible_timeout to 10 13118 1727204079.38621: Set connection var ansible_pipelining to False 13118 1727204079.38628: Set connection var ansible_connection to ssh 13118 1727204079.38646: Set connection var ansible_shell_executable to /bin/sh 13118 1727204079.38656: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204079.38662: Set connection var ansible_shell_type to sh 13118 1727204079.38694: variable 'ansible_shell_executable' from source: unknown 13118 1727204079.38702: variable 'ansible_connection' from source: unknown 13118 1727204079.38709: variable 'ansible_module_compression' from source: unknown 13118 1727204079.38715: variable 'ansible_shell_type' from source: unknown 13118 1727204079.38722: variable 'ansible_shell_executable' from source: unknown 13118 1727204079.38728: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204079.38740: variable 'ansible_pipelining' from source: unknown 13118 1727204079.38751: variable 'ansible_timeout' from source: unknown 13118 1727204079.38759: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204079.38873: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204079.38889: variable 'omit' from source: magic vars 13118 1727204079.38899: starting attempt loop 13118 1727204079.38905: running the handler 13118 1727204079.38992: variable 'ansible_facts' from source: unknown 13118 1727204079.39737: _low_level_execute_command(): starting 13118 1727204079.39751: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204079.40539: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204079.40556: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204079.40572: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204079.40591: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204079.40642: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204079.40655: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204079.40671: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204079.40688: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204079.40699: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204079.40709: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204079.40723: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204079.40740: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204079.40755: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204079.40769: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204079.40780: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204079.40792: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204079.40877: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204079.40893: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204079.40906: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204079.40988: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204079.42661: stdout chunk (state=3): >>>/root <<< 13118 1727204079.42762: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204079.42858: stderr chunk (state=3): >>><<< 13118 1727204079.42870: stdout chunk (state=3): >>><<< 13118 1727204079.42986: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204079.42989: _low_level_execute_command(): starting 13118 1727204079.42992: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204079.4289076-14908-204881174146169 `" && echo ansible-tmp-1727204079.4289076-14908-204881174146169="` echo /root/.ansible/tmp/ansible-tmp-1727204079.4289076-14908-204881174146169 `" ) && sleep 0' 13118 1727204079.43596: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204079.43614: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204079.43632: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204079.43651: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204079.43699: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204079.43715: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204079.43733: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204079.43753: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204079.43768: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204079.43781: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204079.43793: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204079.43807: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204079.43827: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204079.43844: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204079.43856: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204079.43873: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204079.43952: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204079.43976: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204079.43992: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204079.44073: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204079.45926: stdout chunk (state=3): >>>ansible-tmp-1727204079.4289076-14908-204881174146169=/root/.ansible/tmp/ansible-tmp-1727204079.4289076-14908-204881174146169 <<< 13118 1727204079.46042: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204079.46107: stderr chunk (state=3): >>><<< 13118 1727204079.46110: stdout chunk (state=3): >>><<< 13118 1727204079.46119: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204079.4289076-14908-204881174146169=/root/.ansible/tmp/ansible-tmp-1727204079.4289076-14908-204881174146169 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204079.46145: variable 'ansible_module_compression' from source: unknown 13118 1727204079.46193: ANSIBALLZ: Using generic lock for ansible.legacy.systemd 13118 1727204079.46197: ANSIBALLZ: Acquiring lock 13118 1727204079.46199: ANSIBALLZ: Lock acquired: 140051944306976 13118 1727204079.46201: ANSIBALLZ: Creating module 13118 1727204079.71016: ANSIBALLZ: Writing module into payload 13118 1727204079.71153: ANSIBALLZ: Writing module 13118 1727204079.71185: ANSIBALLZ: Renaming module 13118 1727204079.71189: ANSIBALLZ: Done creating module 13118 1727204079.71210: variable 'ansible_facts' from source: unknown 13118 1727204079.71308: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204079.4289076-14908-204881174146169/AnsiballZ_systemd.py 13118 1727204079.71425: Sending initial data 13118 1727204079.71428: Sent initial data (156 bytes) 13118 1727204079.72183: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204079.72189: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204079.72217: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204079.72221: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204079.72223: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204079.72270: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204079.72283: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204079.72340: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204079.74199: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204079.74236: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204079.74271: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmp76q3l54v /root/.ansible/tmp/ansible-tmp-1727204079.4289076-14908-204881174146169/AnsiballZ_systemd.py <<< 13118 1727204079.74306: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204079.76061: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204079.76178: stderr chunk (state=3): >>><<< 13118 1727204079.76182: stdout chunk (state=3): >>><<< 13118 1727204079.76200: done transferring module to remote 13118 1727204079.76209: _low_level_execute_command(): starting 13118 1727204079.76214: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204079.4289076-14908-204881174146169/ /root/.ansible/tmp/ansible-tmp-1727204079.4289076-14908-204881174146169/AnsiballZ_systemd.py && sleep 0' 13118 1727204079.76892: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204079.76950: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204079.76957: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204079.76960: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204079.77016: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204079.78757: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204079.78815: stderr chunk (state=3): >>><<< 13118 1727204079.78818: stdout chunk (state=3): >>><<< 13118 1727204079.78848: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204079.78851: _low_level_execute_command(): starting 13118 1727204079.78854: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204079.4289076-14908-204881174146169/AnsiballZ_systemd.py && sleep 0' 13118 1727204079.79324: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204079.79347: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204079.79361: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204079.79382: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204079.79446: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204079.79465: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204079.79490: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204079.79517: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204079.79529: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204079.79622: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204079.79633: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204079.79645: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204079.79658: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204079.79773: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204079.79785: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204079.79792: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204079.79794: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204079.79796: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204079.79798: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204079.80034: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204080.05519: stdout chunk (state=3): >>> {"name": "NetworkManager", "changed": false, "status": {"Type": "dbus", "ExitType": "main", "Restart": "on-failure", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "10min", "TimeoutStopUSec": "1min 30s", "TimeoutAbortUSec": "1min 30s", "TimeoutStartFailureMode": "terminate", "TimeoutStopFailureMode": "terminate", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "WatchdogUSec": "0", "WatchdogTimestampMonotonic": "0", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "6823", "ControlPID": "0", "BusName": "org.freedesktop.NetworkManager", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "ReloadResult": "success", "CleanResult": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "OOMPolicy": "stop", "ReloadSignal": "1", "ExecMainStartTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ExecMainStartTimestampMonotonic": "319366198", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "6823", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReload": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/NetworkManag<<< 13118 1727204080.05545: stdout chunk (state=3): >>>er.service", "ControlGroupId": "3602", "MemoryCurrent": "6250496", "MemoryAvailable": "infinity", "CPUUsageNSec": "435491000", "TasksCurrent": "3", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "Delegate": "no", "CPUAccounting": "yes", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "MemoryHigh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22342", "IPAccounting": "no", "ManagedOOMSwap": "auto", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "65536", "LimitNOFILESoft": "65536", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "13964", "LimitNPROCSoft": "13964", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "13964", "LimitSIGPENDINGSoft": "13964", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "CoredumpFilter": "0x33", "Nice": "0", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_dac_override cap_kill cap_setgid cap_setuid cap_net_bind_service cap_net_admin cap_net_raw cap_sys_module cap_sys_chroot cap_audit_write", "DynamicUser": "no", "RemoveIPC": "no", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectClock": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectKernelLogs": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "PrivateIPC": "no", "ProtectHome": "read-only", "ProtectSystem": "yes", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "2147483646", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "TimeoutCleanUSec": "infinity", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "ProtectProc": "default", "ProcSubset": "all", "ProtectHostname": "no", "KillMode": "process", "KillSignal": "15", "RestartKillSignal": "15", "FinalKillSignal": "9", "SendSIGKILL": "yes", "SendSIGHUP": "no", "WatchdogSignal": "6", "Id": "NetworkManager.service", "Names": "NetworkManager.service", "Requires": "dbus.socket system.slice sysinit.target", "Wants": "network.target", "BindsTo": "dbus-broker.service", "RequiredBy": "NetworkManager-wait-online.service", "WantedBy": "multi-user.target", "Conflicts": "shutdown.target", "Before": "NetworkManager-wait-online.service cloud-init.service network.target network.service multi-user.target shutdown.target", "After": "systemd-journald.socket network-pre.target dbus-broker.service cloud-init-local.service system.slice basic.target dbus.socket sysinit.target", "Documentation": "\"man:NetworkManager(8)\"", "Description": "Network Manager", "AccessSELinuxContext": "system_u:object_r:NetworkManager_unit_file_t:s0", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/NetworkManager.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Tue 2024-09-24 14:54:30 EDT", "StateChangeTimestampMonotonic": "433536261", "InactiveExitTimestamp": "Tue 2024-09-24 14:52:36 EDT", "InactiveExitTimestampMonotonic": "319366492", "ActiveEnterTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ActiveEnterTimestampMonotonic": "319444795", "ActiveExitTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ActiveExitTimestampMonotonic": "319337881", "InactiveEnterTimestamp": "Tue 2024-09-24 14:52:36 EDT", "InactiveEnterTimestampMonotonic": "319361759", "CanStart": "yes", "CanStop": "yes", "CanReload": "yes", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnSuccessJobMode": "fail", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ConditionTimestampMonotonic": "319362324", "AssertTimestamp": "Tue 2024-09-24 14:52:36 EDT", "AssertTimestampMonotonic": "319362327", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "bc82db972fb14d0fb9ce19d409aedafe", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "NetworkManager", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} <<< 13118 1727204080.07189: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204080.07193: stdout chunk (state=3): >>><<< 13118 1727204080.07199: stderr chunk (state=3): >>><<< 13118 1727204080.07226: _low_level_execute_command() done: rc=0, stdout= {"name": "NetworkManager", "changed": false, "status": {"Type": "dbus", "ExitType": "main", "Restart": "on-failure", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "10min", "TimeoutStopUSec": "1min 30s", "TimeoutAbortUSec": "1min 30s", "TimeoutStartFailureMode": "terminate", "TimeoutStopFailureMode": "terminate", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "WatchdogUSec": "0", "WatchdogTimestampMonotonic": "0", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "6823", "ControlPID": "0", "BusName": "org.freedesktop.NetworkManager", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "ReloadResult": "success", "CleanResult": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "OOMPolicy": "stop", "ReloadSignal": "1", "ExecMainStartTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ExecMainStartTimestampMonotonic": "319366198", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "6823", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReload": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/NetworkManager.service", "ControlGroupId": "3602", "MemoryCurrent": "6250496", "MemoryAvailable": "infinity", "CPUUsageNSec": "435491000", "TasksCurrent": "3", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "Delegate": "no", "CPUAccounting": "yes", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "MemoryHigh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22342", "IPAccounting": "no", "ManagedOOMSwap": "auto", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "65536", "LimitNOFILESoft": "65536", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "13964", "LimitNPROCSoft": "13964", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "13964", "LimitSIGPENDINGSoft": "13964", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "CoredumpFilter": "0x33", "Nice": "0", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_dac_override cap_kill cap_setgid cap_setuid cap_net_bind_service cap_net_admin cap_net_raw cap_sys_module cap_sys_chroot cap_audit_write", "DynamicUser": "no", "RemoveIPC": "no", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectClock": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectKernelLogs": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "PrivateIPC": "no", "ProtectHome": "read-only", "ProtectSystem": "yes", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "2147483646", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "TimeoutCleanUSec": "infinity", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "ProtectProc": "default", "ProcSubset": "all", "ProtectHostname": "no", "KillMode": "process", "KillSignal": "15", "RestartKillSignal": "15", "FinalKillSignal": "9", "SendSIGKILL": "yes", "SendSIGHUP": "no", "WatchdogSignal": "6", "Id": "NetworkManager.service", "Names": "NetworkManager.service", "Requires": "dbus.socket system.slice sysinit.target", "Wants": "network.target", "BindsTo": "dbus-broker.service", "RequiredBy": "NetworkManager-wait-online.service", "WantedBy": "multi-user.target", "Conflicts": "shutdown.target", "Before": "NetworkManager-wait-online.service cloud-init.service network.target network.service multi-user.target shutdown.target", "After": "systemd-journald.socket network-pre.target dbus-broker.service cloud-init-local.service system.slice basic.target dbus.socket sysinit.target", "Documentation": "\"man:NetworkManager(8)\"", "Description": "Network Manager", "AccessSELinuxContext": "system_u:object_r:NetworkManager_unit_file_t:s0", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/NetworkManager.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Tue 2024-09-24 14:54:30 EDT", "StateChangeTimestampMonotonic": "433536261", "InactiveExitTimestamp": "Tue 2024-09-24 14:52:36 EDT", "InactiveExitTimestampMonotonic": "319366492", "ActiveEnterTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ActiveEnterTimestampMonotonic": "319444795", "ActiveExitTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ActiveExitTimestampMonotonic": "319337881", "InactiveEnterTimestamp": "Tue 2024-09-24 14:52:36 EDT", "InactiveEnterTimestampMonotonic": "319361759", "CanStart": "yes", "CanStop": "yes", "CanReload": "yes", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnSuccessJobMode": "fail", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ConditionTimestampMonotonic": "319362324", "AssertTimestamp": "Tue 2024-09-24 14:52:36 EDT", "AssertTimestampMonotonic": "319362327", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "bc82db972fb14d0fb9ce19d409aedafe", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "NetworkManager", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204080.07415: done with _execute_module (ansible.legacy.systemd, {'name': 'NetworkManager', 'state': 'started', 'enabled': True, '_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.systemd', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204079.4289076-14908-204881174146169/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204080.07422: _low_level_execute_command(): starting 13118 1727204080.07427: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204079.4289076-14908-204881174146169/ > /dev/null 2>&1 && sleep 0' 13118 1727204080.10027: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204080.10083: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204080.10093: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204080.10191: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204080.10230: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204080.10240: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204080.10249: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204080.10261: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204080.10272: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204080.10279: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204080.10292: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204080.10300: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204080.10310: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204080.10318: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204080.10324: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204080.10336: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204080.10487: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204080.10511: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204080.10523: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204080.10592: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204080.12482: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204080.12515: stderr chunk (state=3): >>><<< 13118 1727204080.12518: stdout chunk (state=3): >>><<< 13118 1727204080.12538: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204080.12544: handler run complete 13118 1727204080.12607: attempt loop complete, returning result 13118 1727204080.12610: _execute() done 13118 1727204080.12613: dumping result to json 13118 1727204080.12631: done dumping result, returning 13118 1727204080.12643: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Enable and start NetworkManager [0affcd87-79f5-56a3-0a64-000000000032] 13118 1727204080.12648: sending task result for task 0affcd87-79f5-56a3-0a64-000000000032 13118 1727204080.12929: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000032 13118 1727204080.12932: WORKER PROCESS EXITING ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13118 1727204080.12982: no more pending results, returning what we have 13118 1727204080.12985: results queue empty 13118 1727204080.12986: checking for any_errors_fatal 13118 1727204080.12993: done checking for any_errors_fatal 13118 1727204080.12993: checking for max_fail_percentage 13118 1727204080.12995: done checking for max_fail_percentage 13118 1727204080.12996: checking to see if all hosts have failed and the running result is not ok 13118 1727204080.12996: done checking to see if all hosts have failed 13118 1727204080.12997: getting the remaining hosts for this loop 13118 1727204080.12998: done getting the remaining hosts for this loop 13118 1727204080.13002: getting the next task for host managed-node2 13118 1727204080.13008: done getting next task for host managed-node2 13118 1727204080.13012: ^ task is: TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant 13118 1727204080.13014: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=17, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204080.13023: getting variables 13118 1727204080.13025: in VariableManager get_vars() 13118 1727204080.13061: Calling all_inventory to load vars for managed-node2 13118 1727204080.13063: Calling groups_inventory to load vars for managed-node2 13118 1727204080.13069: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204080.13078: Calling all_plugins_play to load vars for managed-node2 13118 1727204080.13081: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204080.13083: Calling groups_plugins_play to load vars for managed-node2 13118 1727204080.15278: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204080.17681: done with get_vars() 13118 1727204080.17712: done getting variables 13118 1727204080.17785: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable and start wpa_supplicant] ***** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:133 Tuesday 24 September 2024 14:54:40 -0400 (0:00:00.911) 0:00:17.478 ***** 13118 1727204080.17826: entering _queue_task() for managed-node2/service 13118 1727204080.18171: worker is 1 (out of 1 available) 13118 1727204080.18187: exiting _queue_task() for managed-node2/service 13118 1727204080.18201: done queuing things up, now waiting for results queue to drain 13118 1727204080.18202: waiting for pending results... 13118 1727204080.18496: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant 13118 1727204080.18659: in run() - task 0affcd87-79f5-56a3-0a64-000000000033 13118 1727204080.18683: variable 'ansible_search_path' from source: unknown 13118 1727204080.18691: variable 'ansible_search_path' from source: unknown 13118 1727204080.18738: calling self._execute() 13118 1727204080.18839: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204080.18852: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204080.18873: variable 'omit' from source: magic vars 13118 1727204080.19390: variable 'ansible_distribution_major_version' from source: facts 13118 1727204080.19408: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204080.19529: variable 'network_provider' from source: set_fact 13118 1727204080.19541: Evaluated conditional (network_provider == "nm"): True 13118 1727204080.19738: variable '__network_wpa_supplicant_required' from source: role '' defaults 13118 1727204080.19951: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13118 1727204080.20300: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204080.24004: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204080.24108: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204080.24299: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204080.24348: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204080.24501: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204080.24707: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204080.24748: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204080.24784: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204080.24920: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204080.24945: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204080.25069: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204080.25100: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204080.25262: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204080.25311: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204080.25336: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204080.25474: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204080.25503: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204080.25537: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204080.25602: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204080.25686: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204080.25968: variable 'network_connections' from source: task vars 13118 1727204080.26113: variable 'controller_profile' from source: play vars 13118 1727204080.26197: variable 'controller_profile' from source: play vars 13118 1727204080.26327: variable 'controller_device' from source: play vars 13118 1727204080.26404: variable 'controller_device' from source: play vars 13118 1727204080.26538: variable 'port1_profile' from source: play vars 13118 1727204080.26605: variable 'port1_profile' from source: play vars 13118 1727204080.26618: variable 'dhcp_interface1' from source: play vars 13118 1727204080.26688: variable 'dhcp_interface1' from source: play vars 13118 1727204080.26758: variable 'controller_profile' from source: play vars 13118 1727204080.26823: variable 'controller_profile' from source: play vars 13118 1727204080.26977: variable 'port2_profile' from source: play vars 13118 1727204080.27041: variable 'port2_profile' from source: play vars 13118 1727204080.27054: variable 'dhcp_interface2' from source: play vars 13118 1727204080.27237: variable 'dhcp_interface2' from source: play vars 13118 1727204080.27249: variable 'controller_profile' from source: play vars 13118 1727204080.27318: variable 'controller_profile' from source: play vars 13118 1727204080.27508: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204080.27709: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204080.27757: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204080.27793: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204080.27826: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204080.27881: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204080.27906: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204080.27940: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204080.27986: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204080.28046: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204080.28331: variable 'network_connections' from source: task vars 13118 1727204080.28343: variable 'controller_profile' from source: play vars 13118 1727204080.28416: variable 'controller_profile' from source: play vars 13118 1727204080.28428: variable 'controller_device' from source: play vars 13118 1727204080.28495: variable 'controller_device' from source: play vars 13118 1727204080.28511: variable 'port1_profile' from source: play vars 13118 1727204080.28576: variable 'port1_profile' from source: play vars 13118 1727204080.28588: variable 'dhcp_interface1' from source: play vars 13118 1727204080.28656: variable 'dhcp_interface1' from source: play vars 13118 1727204080.28672: variable 'controller_profile' from source: play vars 13118 1727204080.28741: variable 'controller_profile' from source: play vars 13118 1727204080.28754: variable 'port2_profile' from source: play vars 13118 1727204080.28819: variable 'port2_profile' from source: play vars 13118 1727204080.28838: variable 'dhcp_interface2' from source: play vars 13118 1727204080.28901: variable 'dhcp_interface2' from source: play vars 13118 1727204080.28912: variable 'controller_profile' from source: play vars 13118 1727204080.28982: variable 'controller_profile' from source: play vars 13118 1727204080.29032: Evaluated conditional (__network_wpa_supplicant_required): False 13118 1727204080.29042: when evaluation is False, skipping this task 13118 1727204080.29051: _execute() done 13118 1727204080.29058: dumping result to json 13118 1727204080.29066: done dumping result, returning 13118 1727204080.29077: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant [0affcd87-79f5-56a3-0a64-000000000033] 13118 1727204080.29086: sending task result for task 0affcd87-79f5-56a3-0a64-000000000033 skipping: [managed-node2] => { "changed": false, "false_condition": "__network_wpa_supplicant_required", "skip_reason": "Conditional result was False" } 13118 1727204080.29288: no more pending results, returning what we have 13118 1727204080.29292: results queue empty 13118 1727204080.29293: checking for any_errors_fatal 13118 1727204080.29312: done checking for any_errors_fatal 13118 1727204080.29312: checking for max_fail_percentage 13118 1727204080.29314: done checking for max_fail_percentage 13118 1727204080.29315: checking to see if all hosts have failed and the running result is not ok 13118 1727204080.29316: done checking to see if all hosts have failed 13118 1727204080.29316: getting the remaining hosts for this loop 13118 1727204080.29318: done getting the remaining hosts for this loop 13118 1727204080.29322: getting the next task for host managed-node2 13118 1727204080.29332: done getting next task for host managed-node2 13118 1727204080.29337: ^ task is: TASK: fedora.linux_system_roles.network : Enable network service 13118 1727204080.29340: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204080.29356: getting variables 13118 1727204080.29358: in VariableManager get_vars() 13118 1727204080.29406: Calling all_inventory to load vars for managed-node2 13118 1727204080.29409: Calling groups_inventory to load vars for managed-node2 13118 1727204080.29411: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204080.29422: Calling all_plugins_play to load vars for managed-node2 13118 1727204080.29426: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204080.29432: Calling groups_plugins_play to load vars for managed-node2 13118 1727204080.30537: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000033 13118 1727204080.30541: WORKER PROCESS EXITING 13118 1727204080.31202: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204080.32993: done with get_vars() 13118 1727204080.33021: done getting variables 13118 1727204080.33089: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable network service] ************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:142 Tuesday 24 September 2024 14:54:40 -0400 (0:00:00.152) 0:00:17.631 ***** 13118 1727204080.33123: entering _queue_task() for managed-node2/service 13118 1727204080.33456: worker is 1 (out of 1 available) 13118 1727204080.33471: exiting _queue_task() for managed-node2/service 13118 1727204080.33486: done queuing things up, now waiting for results queue to drain 13118 1727204080.33487: waiting for pending results... 13118 1727204080.33787: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Enable network service 13118 1727204080.33941: in run() - task 0affcd87-79f5-56a3-0a64-000000000034 13118 1727204080.33961: variable 'ansible_search_path' from source: unknown 13118 1727204080.33970: variable 'ansible_search_path' from source: unknown 13118 1727204080.34013: calling self._execute() 13118 1727204080.34107: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204080.34118: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204080.34133: variable 'omit' from source: magic vars 13118 1727204080.34552: variable 'ansible_distribution_major_version' from source: facts 13118 1727204080.34572: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204080.34699: variable 'network_provider' from source: set_fact 13118 1727204080.34708: Evaluated conditional (network_provider == "initscripts"): False 13118 1727204080.34714: when evaluation is False, skipping this task 13118 1727204080.34723: _execute() done 13118 1727204080.34732: dumping result to json 13118 1727204080.34741: done dumping result, returning 13118 1727204080.34751: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Enable network service [0affcd87-79f5-56a3-0a64-000000000034] 13118 1727204080.34762: sending task result for task 0affcd87-79f5-56a3-0a64-000000000034 13118 1727204080.34886: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000034 13118 1727204080.34893: WORKER PROCESS EXITING skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13118 1727204080.34947: no more pending results, returning what we have 13118 1727204080.34951: results queue empty 13118 1727204080.34952: checking for any_errors_fatal 13118 1727204080.34959: done checking for any_errors_fatal 13118 1727204080.34960: checking for max_fail_percentage 13118 1727204080.34962: done checking for max_fail_percentage 13118 1727204080.34963: checking to see if all hosts have failed and the running result is not ok 13118 1727204080.34965: done checking to see if all hosts have failed 13118 1727204080.34966: getting the remaining hosts for this loop 13118 1727204080.34967: done getting the remaining hosts for this loop 13118 1727204080.34971: getting the next task for host managed-node2 13118 1727204080.34979: done getting next task for host managed-node2 13118 1727204080.34983: ^ task is: TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present 13118 1727204080.34985: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=19, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204080.35001: getting variables 13118 1727204080.35003: in VariableManager get_vars() 13118 1727204080.35050: Calling all_inventory to load vars for managed-node2 13118 1727204080.35053: Calling groups_inventory to load vars for managed-node2 13118 1727204080.35055: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204080.35070: Calling all_plugins_play to load vars for managed-node2 13118 1727204080.35073: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204080.35076: Calling groups_plugins_play to load vars for managed-node2 13118 1727204080.37452: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204080.39153: done with get_vars() 13118 1727204080.39186: done getting variables 13118 1727204080.39250: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Ensure initscripts network file dependency is present] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:150 Tuesday 24 September 2024 14:54:40 -0400 (0:00:00.061) 0:00:17.692 ***** 13118 1727204080.39289: entering _queue_task() for managed-node2/copy 13118 1727204080.39683: worker is 1 (out of 1 available) 13118 1727204080.39695: exiting _queue_task() for managed-node2/copy 13118 1727204080.39708: done queuing things up, now waiting for results queue to drain 13118 1727204080.39709: waiting for pending results... 13118 1727204080.40002: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present 13118 1727204080.40151: in run() - task 0affcd87-79f5-56a3-0a64-000000000035 13118 1727204080.40176: variable 'ansible_search_path' from source: unknown 13118 1727204080.40186: variable 'ansible_search_path' from source: unknown 13118 1727204080.40228: calling self._execute() 13118 1727204080.40332: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204080.40345: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204080.40359: variable 'omit' from source: magic vars 13118 1727204080.40752: variable 'ansible_distribution_major_version' from source: facts 13118 1727204080.40774: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204080.40900: variable 'network_provider' from source: set_fact 13118 1727204080.40915: Evaluated conditional (network_provider == "initscripts"): False 13118 1727204080.40923: when evaluation is False, skipping this task 13118 1727204080.40932: _execute() done 13118 1727204080.40939: dumping result to json 13118 1727204080.40946: done dumping result, returning 13118 1727204080.40956: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present [0affcd87-79f5-56a3-0a64-000000000035] 13118 1727204080.40968: sending task result for task 0affcd87-79f5-56a3-0a64-000000000035 skipping: [managed-node2] => { "changed": false, "false_condition": "network_provider == \"initscripts\"", "skip_reason": "Conditional result was False" } 13118 1727204080.41124: no more pending results, returning what we have 13118 1727204080.41132: results queue empty 13118 1727204080.41133: checking for any_errors_fatal 13118 1727204080.41140: done checking for any_errors_fatal 13118 1727204080.41141: checking for max_fail_percentage 13118 1727204080.41143: done checking for max_fail_percentage 13118 1727204080.41144: checking to see if all hosts have failed and the running result is not ok 13118 1727204080.41145: done checking to see if all hosts have failed 13118 1727204080.41145: getting the remaining hosts for this loop 13118 1727204080.41147: done getting the remaining hosts for this loop 13118 1727204080.41151: getting the next task for host managed-node2 13118 1727204080.41159: done getting next task for host managed-node2 13118 1727204080.41163: ^ task is: TASK: fedora.linux_system_roles.network : Configure networking connection profiles 13118 1727204080.41168: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=20, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204080.41184: getting variables 13118 1727204080.41186: in VariableManager get_vars() 13118 1727204080.41285: Calling all_inventory to load vars for managed-node2 13118 1727204080.41288: Calling groups_inventory to load vars for managed-node2 13118 1727204080.41290: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204080.41330: Calling all_plugins_play to load vars for managed-node2 13118 1727204080.41336: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204080.41340: Calling groups_plugins_play to load vars for managed-node2 13118 1727204080.43540: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000035 13118 1727204080.43545: WORKER PROCESS EXITING 13118 1727204080.44141: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204080.46040: done with get_vars() 13118 1727204080.46063: done getting variables TASK [fedora.linux_system_roles.network : Configure networking connection profiles] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:159 Tuesday 24 September 2024 14:54:40 -0400 (0:00:00.068) 0:00:17.761 ***** 13118 1727204080.46153: entering _queue_task() for managed-node2/fedora.linux_system_roles.network_connections 13118 1727204080.46154: Creating lock for fedora.linux_system_roles.network_connections 13118 1727204080.46476: worker is 1 (out of 1 available) 13118 1727204080.46489: exiting _queue_task() for managed-node2/fedora.linux_system_roles.network_connections 13118 1727204080.46502: done queuing things up, now waiting for results queue to drain 13118 1727204080.46503: waiting for pending results... 13118 1727204080.46817: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Configure networking connection profiles 13118 1727204080.46954: in run() - task 0affcd87-79f5-56a3-0a64-000000000036 13118 1727204080.46977: variable 'ansible_search_path' from source: unknown 13118 1727204080.46985: variable 'ansible_search_path' from source: unknown 13118 1727204080.47023: calling self._execute() 13118 1727204080.47121: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204080.47136: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204080.47149: variable 'omit' from source: magic vars 13118 1727204080.47553: variable 'ansible_distribution_major_version' from source: facts 13118 1727204080.47576: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204080.47588: variable 'omit' from source: magic vars 13118 1727204080.47651: variable 'omit' from source: magic vars 13118 1727204080.47834: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204080.50284: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204080.50365: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204080.50407: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204080.50452: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204080.50485: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204080.50582: variable 'network_provider' from source: set_fact 13118 1727204080.50726: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204080.50785: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204080.50819: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204080.50875: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204080.50895: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204080.50984: variable 'omit' from source: magic vars 13118 1727204080.51109: variable 'omit' from source: magic vars 13118 1727204080.51221: variable 'network_connections' from source: task vars 13118 1727204080.51240: variable 'controller_profile' from source: play vars 13118 1727204080.51311: variable 'controller_profile' from source: play vars 13118 1727204080.51324: variable 'controller_device' from source: play vars 13118 1727204080.51391: variable 'controller_device' from source: play vars 13118 1727204080.51410: variable 'port1_profile' from source: play vars 13118 1727204080.51475: variable 'port1_profile' from source: play vars 13118 1727204080.51487: variable 'dhcp_interface1' from source: play vars 13118 1727204080.51553: variable 'dhcp_interface1' from source: play vars 13118 1727204080.51567: variable 'controller_profile' from source: play vars 13118 1727204080.51636: variable 'controller_profile' from source: play vars 13118 1727204080.51649: variable 'port2_profile' from source: play vars 13118 1727204080.51714: variable 'port2_profile' from source: play vars 13118 1727204080.51733: variable 'dhcp_interface2' from source: play vars 13118 1727204080.51798: variable 'dhcp_interface2' from source: play vars 13118 1727204080.51810: variable 'controller_profile' from source: play vars 13118 1727204080.51881: variable 'controller_profile' from source: play vars 13118 1727204080.52081: variable 'omit' from source: magic vars 13118 1727204080.52094: variable '__lsr_ansible_managed' from source: task vars 13118 1727204080.52160: variable '__lsr_ansible_managed' from source: task vars 13118 1727204080.52351: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup 13118 1727204080.52593: Loaded config def from plugin (lookup/template) 13118 1727204080.52605: Loading LookupModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup/template.py 13118 1727204080.52640: File lookup term: get_ansible_managed.j2 13118 1727204080.52648: variable 'ansible_search_path' from source: unknown 13118 1727204080.52658: evaluation_path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks 13118 1727204080.52679: search_path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/templates/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/templates/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/templates/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/templates/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/get_ansible_managed.j2 13118 1727204080.52700: variable 'ansible_search_path' from source: unknown 13118 1727204080.59616: variable 'ansible_managed' from source: unknown 13118 1727204080.59885: variable 'omit' from source: magic vars 13118 1727204080.59921: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204080.59958: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204080.59987: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204080.60011: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204080.60025: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204080.60059: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204080.60074: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204080.60083: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204080.60189: Set connection var ansible_timeout to 10 13118 1727204080.60207: Set connection var ansible_pipelining to False 13118 1727204080.60213: Set connection var ansible_connection to ssh 13118 1727204080.60223: Set connection var ansible_shell_executable to /bin/sh 13118 1727204080.60235: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204080.60296: Set connection var ansible_shell_type to sh 13118 1727204080.60326: variable 'ansible_shell_executable' from source: unknown 13118 1727204080.60376: variable 'ansible_connection' from source: unknown 13118 1727204080.60385: variable 'ansible_module_compression' from source: unknown 13118 1727204080.60393: variable 'ansible_shell_type' from source: unknown 13118 1727204080.60403: variable 'ansible_shell_executable' from source: unknown 13118 1727204080.60411: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204080.60419: variable 'ansible_pipelining' from source: unknown 13118 1727204080.60425: variable 'ansible_timeout' from source: unknown 13118 1727204080.60436: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204080.60685: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204080.60700: variable 'omit' from source: magic vars 13118 1727204080.60711: starting attempt loop 13118 1727204080.60719: running the handler 13118 1727204080.60743: _low_level_execute_command(): starting 13118 1727204080.60756: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204080.61520: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204080.61541: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204080.61557: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204080.61584: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204080.61636: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204080.61650: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204080.61667: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204080.61686: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204080.61697: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204080.61710: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204080.61727: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204080.61746: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204080.61866: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204080.61880: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204080.61892: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204080.61905: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204080.61987: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204080.62004: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204080.62020: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204080.62103: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204080.63799: stdout chunk (state=3): >>>/root <<< 13118 1727204080.63882: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204080.63974: stderr chunk (state=3): >>><<< 13118 1727204080.63977: stdout chunk (state=3): >>><<< 13118 1727204080.64097: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204080.64100: _low_level_execute_command(): starting 13118 1727204080.64103: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204080.6399798-14954-197686272481751 `" && echo ansible-tmp-1727204080.6399798-14954-197686272481751="` echo /root/.ansible/tmp/ansible-tmp-1727204080.6399798-14954-197686272481751 `" ) && sleep 0' 13118 1727204080.65462: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204080.65467: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204080.65501: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204080.65505: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204080.65507: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204080.65571: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204080.65652: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204080.65725: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204080.67618: stdout chunk (state=3): >>>ansible-tmp-1727204080.6399798-14954-197686272481751=/root/.ansible/tmp/ansible-tmp-1727204080.6399798-14954-197686272481751 <<< 13118 1727204080.67813: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204080.67816: stdout chunk (state=3): >>><<< 13118 1727204080.67819: stderr chunk (state=3): >>><<< 13118 1727204080.68270: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204080.6399798-14954-197686272481751=/root/.ansible/tmp/ansible-tmp-1727204080.6399798-14954-197686272481751 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204080.68281: variable 'ansible_module_compression' from source: unknown 13118 1727204080.68284: ANSIBALLZ: Using lock for fedora.linux_system_roles.network_connections 13118 1727204080.68286: ANSIBALLZ: Acquiring lock 13118 1727204080.68288: ANSIBALLZ: Lock acquired: 140051940594976 13118 1727204080.68290: ANSIBALLZ: Creating module 13118 1727204080.91552: ANSIBALLZ: Writing module into payload 13118 1727204080.91953: ANSIBALLZ: Writing module 13118 1727204080.91979: ANSIBALLZ: Renaming module 13118 1727204080.91983: ANSIBALLZ: Done creating module 13118 1727204080.92002: variable 'ansible_facts' from source: unknown 13118 1727204080.92069: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204080.6399798-14954-197686272481751/AnsiballZ_network_connections.py 13118 1727204080.92175: Sending initial data 13118 1727204080.92178: Sent initial data (168 bytes) 13118 1727204080.92848: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204080.92854: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204080.92886: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204080.92901: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204080.92911: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204080.92960: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204080.92976: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204080.93029: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204080.94839: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204080.94877: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204080.94912: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmp581jmrlv /root/.ansible/tmp/ansible-tmp-1727204080.6399798-14954-197686272481751/AnsiballZ_network_connections.py <<< 13118 1727204080.94946: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204080.96089: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204080.96202: stderr chunk (state=3): >>><<< 13118 1727204080.96206: stdout chunk (state=3): >>><<< 13118 1727204080.96232: done transferring module to remote 13118 1727204080.96240: _low_level_execute_command(): starting 13118 1727204080.96245: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204080.6399798-14954-197686272481751/ /root/.ansible/tmp/ansible-tmp-1727204080.6399798-14954-197686272481751/AnsiballZ_network_connections.py && sleep 0' 13118 1727204080.96724: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204080.96740: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204080.96760: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204080.96778: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204080.96835: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204080.96841: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204080.96893: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204080.98650: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204080.98704: stderr chunk (state=3): >>><<< 13118 1727204080.98709: stdout chunk (state=3): >>><<< 13118 1727204080.98722: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204080.98725: _low_level_execute_command(): starting 13118 1727204080.98730: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204080.6399798-14954-197686272481751/AnsiballZ_network_connections.py && sleep 0' 13118 1727204080.99178: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204080.99183: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204080.99217: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204080.99229: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204080.99245: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204080.99294: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204080.99306: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204080.99358: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204081.45638: stdout chunk (state=3): >>> {"changed": true, "warnings": [], "stderr": "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, debabe6d-f673-439e-9f14-25954ecc55a3\n[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 382c36b9-9f8f-4077-be2c-b77f342639e8\n[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 161d3351-eab4-440b-a3b9-ee4281b1f747\n[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, debabe6d-f673-439e-9f14-25954ecc55a3 (is-modified)\n[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 382c36b9-9f8f-4077-be2c-b77f342639e8 (not-active)\n[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 161d3351-eab4-440b-a3b9-ee4281b1f747 (not-active)\n", "_invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0", "state": "up", "type": "bond", "interface_name": "nm-bond", "bond": {"mode": "active-backup", "miimon": 110}, "ip": {"route_metric4": 65535}}, {"name": "bond0.0", "state": "up", "type": "ethernet", "interface_name": "test1", "controller": "bond0"}, {"name": "bond0.1", "state": "up", "type": "ethernet", "interface_name": "test2", "controller": "bond0"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}, "invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0", "state": "up", "type": "bond", "interface_name": "nm-bond", "bond": {"mode": "active-backup", "miimon": 110}, "ip": {"route_metric4": 65535}}, {"name": "bond0.0", "state": "up", "type": "ethernet", "interface_name": "test1", "controller": "bond0"}, {"name": "bond0.1", "state": "up", "type": "ethernet", "interface_name": "test2", "controller": "bond0"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}} <<< 13118 1727204081.48076: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204081.48080: stderr chunk (state=3): >>><<< 13118 1727204081.48083: stdout chunk (state=3): >>><<< 13118 1727204081.48105: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "warnings": [], "stderr": "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, debabe6d-f673-439e-9f14-25954ecc55a3\n[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 382c36b9-9f8f-4077-be2c-b77f342639e8\n[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 161d3351-eab4-440b-a3b9-ee4281b1f747\n[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, debabe6d-f673-439e-9f14-25954ecc55a3 (is-modified)\n[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 382c36b9-9f8f-4077-be2c-b77f342639e8 (not-active)\n[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 161d3351-eab4-440b-a3b9-ee4281b1f747 (not-active)\n", "_invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0", "state": "up", "type": "bond", "interface_name": "nm-bond", "bond": {"mode": "active-backup", "miimon": 110}, "ip": {"route_metric4": 65535}}, {"name": "bond0.0", "state": "up", "type": "ethernet", "interface_name": "test1", "controller": "bond0"}, {"name": "bond0.1", "state": "up", "type": "ethernet", "interface_name": "test2", "controller": "bond0"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}, "invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0", "state": "up", "type": "bond", "interface_name": "nm-bond", "bond": {"mode": "active-backup", "miimon": 110}, "ip": {"route_metric4": 65535}}, {"name": "bond0.0", "state": "up", "type": "ethernet", "interface_name": "test1", "controller": "bond0"}, {"name": "bond0.1", "state": "up", "type": "ethernet", "interface_name": "test2", "controller": "bond0"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204081.48159: done with _execute_module (fedora.linux_system_roles.network_connections, {'provider': 'nm', 'connections': [{'name': 'bond0', 'state': 'up', 'type': 'bond', 'interface_name': 'nm-bond', 'bond': {'mode': 'active-backup', 'miimon': 110}, 'ip': {'route_metric4': 65535}}, {'name': 'bond0.0', 'state': 'up', 'type': 'ethernet', 'interface_name': 'test1', 'controller': 'bond0'}, {'name': 'bond0.1', 'state': 'up', 'type': 'ethernet', 'interface_name': 'test2', 'controller': 'bond0'}], '__header': '#\n# Ansible managed\n#\n# system_role:network\n', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'fedora.linux_system_roles.network_connections', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204080.6399798-14954-197686272481751/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204081.48167: _low_level_execute_command(): starting 13118 1727204081.48174: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204080.6399798-14954-197686272481751/ > /dev/null 2>&1 && sleep 0' 13118 1727204081.49637: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204081.49641: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204081.49719: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204081.49728: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204081.49796: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204081.49800: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204081.49940: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204081.50016: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204081.50120: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204081.51971: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204081.52047: stderr chunk (state=3): >>><<< 13118 1727204081.52050: stdout chunk (state=3): >>><<< 13118 1727204081.52060: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204081.52069: handler run complete 13118 1727204081.52110: attempt loop complete, returning result 13118 1727204081.52113: _execute() done 13118 1727204081.52116: dumping result to json 13118 1727204081.52123: done dumping result, returning 13118 1727204081.52135: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Configure networking connection profiles [0affcd87-79f5-56a3-0a64-000000000036] 13118 1727204081.52137: sending task result for task 0affcd87-79f5-56a3-0a64-000000000036 13118 1727204081.52266: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000036 13118 1727204081.52269: WORKER PROCESS EXITING changed: [managed-node2] => { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "bond": { "miimon": 110, "mode": "active-backup" }, "interface_name": "nm-bond", "ip": { "route_metric4": 65535 }, "name": "bond0", "state": "up", "type": "bond" }, { "controller": "bond0", "interface_name": "test1", "name": "bond0.0", "state": "up", "type": "ethernet" }, { "controller": "bond0", "interface_name": "test2", "name": "bond0.1", "state": "up", "type": "ethernet" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true } STDERR: [007] #0, state:up persistent_state:present, 'bond0': add connection bond0, debabe6d-f673-439e-9f14-25954ecc55a3 [008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 382c36b9-9f8f-4077-be2c-b77f342639e8 [009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 161d3351-eab4-440b-a3b9-ee4281b1f747 [010] #0, state:up persistent_state:present, 'bond0': up connection bond0, debabe6d-f673-439e-9f14-25954ecc55a3 (is-modified) [011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 382c36b9-9f8f-4077-be2c-b77f342639e8 (not-active) [012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 161d3351-eab4-440b-a3b9-ee4281b1f747 (not-active) 13118 1727204081.52399: no more pending results, returning what we have 13118 1727204081.52402: results queue empty 13118 1727204081.52403: checking for any_errors_fatal 13118 1727204081.52409: done checking for any_errors_fatal 13118 1727204081.52410: checking for max_fail_percentage 13118 1727204081.52412: done checking for max_fail_percentage 13118 1727204081.52412: checking to see if all hosts have failed and the running result is not ok 13118 1727204081.52413: done checking to see if all hosts have failed 13118 1727204081.52414: getting the remaining hosts for this loop 13118 1727204081.52415: done getting the remaining hosts for this loop 13118 1727204081.52419: getting the next task for host managed-node2 13118 1727204081.52425: done getting next task for host managed-node2 13118 1727204081.52431: ^ task is: TASK: fedora.linux_system_roles.network : Configure networking state 13118 1727204081.52433: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=21, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204081.52443: getting variables 13118 1727204081.52445: in VariableManager get_vars() 13118 1727204081.52485: Calling all_inventory to load vars for managed-node2 13118 1727204081.52488: Calling groups_inventory to load vars for managed-node2 13118 1727204081.52490: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204081.52499: Calling all_plugins_play to load vars for managed-node2 13118 1727204081.52501: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204081.52504: Calling groups_plugins_play to load vars for managed-node2 13118 1727204081.54997: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204081.58122: done with get_vars() 13118 1727204081.58159: done getting variables TASK [fedora.linux_system_roles.network : Configure networking state] ********** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:171 Tuesday 24 September 2024 14:54:41 -0400 (0:00:01.123) 0:00:18.885 ***** 13118 1727204081.58550: entering _queue_task() for managed-node2/fedora.linux_system_roles.network_state 13118 1727204081.58552: Creating lock for fedora.linux_system_roles.network_state 13118 1727204081.59097: worker is 1 (out of 1 available) 13118 1727204081.59112: exiting _queue_task() for managed-node2/fedora.linux_system_roles.network_state 13118 1727204081.59126: done queuing things up, now waiting for results queue to drain 13118 1727204081.59127: waiting for pending results... 13118 1727204081.60039: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Configure networking state 13118 1727204081.60322: in run() - task 0affcd87-79f5-56a3-0a64-000000000037 13118 1727204081.60399: variable 'ansible_search_path' from source: unknown 13118 1727204081.60426: variable 'ansible_search_path' from source: unknown 13118 1727204081.61206: calling self._execute() 13118 1727204081.61815: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204081.61827: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204081.61840: variable 'omit' from source: magic vars 13118 1727204081.62201: variable 'ansible_distribution_major_version' from source: facts 13118 1727204081.62257: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204081.62398: variable 'network_state' from source: role '' defaults 13118 1727204081.62415: Evaluated conditional (network_state != {}): False 13118 1727204081.62422: when evaluation is False, skipping this task 13118 1727204081.62429: _execute() done 13118 1727204081.62435: dumping result to json 13118 1727204081.62443: done dumping result, returning 13118 1727204081.62454: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Configure networking state [0affcd87-79f5-56a3-0a64-000000000037] 13118 1727204081.62466: sending task result for task 0affcd87-79f5-56a3-0a64-000000000037 13118 1727204081.62574: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000037 13118 1727204081.62582: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13118 1727204081.62651: no more pending results, returning what we have 13118 1727204081.62656: results queue empty 13118 1727204081.62657: checking for any_errors_fatal 13118 1727204081.62669: done checking for any_errors_fatal 13118 1727204081.62669: checking for max_fail_percentage 13118 1727204081.62671: done checking for max_fail_percentage 13118 1727204081.62672: checking to see if all hosts have failed and the running result is not ok 13118 1727204081.62672: done checking to see if all hosts have failed 13118 1727204081.62673: getting the remaining hosts for this loop 13118 1727204081.62674: done getting the remaining hosts for this loop 13118 1727204081.62678: getting the next task for host managed-node2 13118 1727204081.62684: done getting next task for host managed-node2 13118 1727204081.62688: ^ task is: TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections 13118 1727204081.62691: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=22, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204081.62706: getting variables 13118 1727204081.62707: in VariableManager get_vars() 13118 1727204081.62750: Calling all_inventory to load vars for managed-node2 13118 1727204081.62753: Calling groups_inventory to load vars for managed-node2 13118 1727204081.62755: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204081.62768: Calling all_plugins_play to load vars for managed-node2 13118 1727204081.62771: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204081.62774: Calling groups_plugins_play to load vars for managed-node2 13118 1727204081.65670: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204081.68742: done with get_vars() 13118 1727204081.68840: done getting variables 13118 1727204081.69023: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show stderr messages for the network_connections] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:177 Tuesday 24 September 2024 14:54:41 -0400 (0:00:00.105) 0:00:18.990 ***** 13118 1727204081.69066: entering _queue_task() for managed-node2/debug 13118 1727204081.69858: worker is 1 (out of 1 available) 13118 1727204081.69874: exiting _queue_task() for managed-node2/debug 13118 1727204081.69887: done queuing things up, now waiting for results queue to drain 13118 1727204081.69888: waiting for pending results... 13118 1727204081.70646: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections 13118 1727204081.70907: in run() - task 0affcd87-79f5-56a3-0a64-000000000038 13118 1727204081.70989: variable 'ansible_search_path' from source: unknown 13118 1727204081.71049: variable 'ansible_search_path' from source: unknown 13118 1727204081.71093: calling self._execute() 13118 1727204081.71191: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204081.71203: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204081.71220: variable 'omit' from source: magic vars 13118 1727204081.71602: variable 'ansible_distribution_major_version' from source: facts 13118 1727204081.71621: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204081.71631: variable 'omit' from source: magic vars 13118 1727204081.71687: variable 'omit' from source: magic vars 13118 1727204081.71729: variable 'omit' from source: magic vars 13118 1727204081.71772: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204081.71814: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204081.71836: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204081.71857: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204081.71873: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204081.71902: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204081.71910: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204081.71920: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204081.72011: Set connection var ansible_timeout to 10 13118 1727204081.72033: Set connection var ansible_pipelining to False 13118 1727204081.72039: Set connection var ansible_connection to ssh 13118 1727204081.72048: Set connection var ansible_shell_executable to /bin/sh 13118 1727204081.72056: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204081.72061: Set connection var ansible_shell_type to sh 13118 1727204081.72085: variable 'ansible_shell_executable' from source: unknown 13118 1727204081.72092: variable 'ansible_connection' from source: unknown 13118 1727204081.72097: variable 'ansible_module_compression' from source: unknown 13118 1727204081.72102: variable 'ansible_shell_type' from source: unknown 13118 1727204081.72108: variable 'ansible_shell_executable' from source: unknown 13118 1727204081.72113: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204081.72118: variable 'ansible_pipelining' from source: unknown 13118 1727204081.72124: variable 'ansible_timeout' from source: unknown 13118 1727204081.72135: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204081.72261: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204081.72279: variable 'omit' from source: magic vars 13118 1727204081.72287: starting attempt loop 13118 1727204081.72293: running the handler 13118 1727204081.72418: variable '__network_connections_result' from source: set_fact 13118 1727204081.72493: handler run complete 13118 1727204081.72514: attempt loop complete, returning result 13118 1727204081.72520: _execute() done 13118 1727204081.72529: dumping result to json 13118 1727204081.72537: done dumping result, returning 13118 1727204081.72547: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections [0affcd87-79f5-56a3-0a64-000000000038] 13118 1727204081.72556: sending task result for task 0affcd87-79f5-56a3-0a64-000000000038 13118 1727204081.72662: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000038 13118 1727204081.72675: WORKER PROCESS EXITING ok: [managed-node2] => { "__network_connections_result.stderr_lines": [ "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, debabe6d-f673-439e-9f14-25954ecc55a3", "[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 382c36b9-9f8f-4077-be2c-b77f342639e8", "[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 161d3351-eab4-440b-a3b9-ee4281b1f747", "[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, debabe6d-f673-439e-9f14-25954ecc55a3 (is-modified)", "[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 382c36b9-9f8f-4077-be2c-b77f342639e8 (not-active)", "[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 161d3351-eab4-440b-a3b9-ee4281b1f747 (not-active)" ] } 13118 1727204081.72740: no more pending results, returning what we have 13118 1727204081.72745: results queue empty 13118 1727204081.72746: checking for any_errors_fatal 13118 1727204081.72752: done checking for any_errors_fatal 13118 1727204081.72753: checking for max_fail_percentage 13118 1727204081.72755: done checking for max_fail_percentage 13118 1727204081.72756: checking to see if all hosts have failed and the running result is not ok 13118 1727204081.72757: done checking to see if all hosts have failed 13118 1727204081.72757: getting the remaining hosts for this loop 13118 1727204081.72759: done getting the remaining hosts for this loop 13118 1727204081.72765: getting the next task for host managed-node2 13118 1727204081.72772: done getting next task for host managed-node2 13118 1727204081.72777: ^ task is: TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections 13118 1727204081.72780: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=23, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204081.72791: getting variables 13118 1727204081.72793: in VariableManager get_vars() 13118 1727204081.72835: Calling all_inventory to load vars for managed-node2 13118 1727204081.72838: Calling groups_inventory to load vars for managed-node2 13118 1727204081.72840: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204081.72851: Calling all_plugins_play to load vars for managed-node2 13118 1727204081.72853: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204081.72856: Calling groups_plugins_play to load vars for managed-node2 13118 1727204081.74606: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204081.76396: done with get_vars() 13118 1727204081.76446: done getting variables 13118 1727204081.76508: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show debug messages for the network_connections] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:181 Tuesday 24 September 2024 14:54:41 -0400 (0:00:00.074) 0:00:19.065 ***** 13118 1727204081.76555: entering _queue_task() for managed-node2/debug 13118 1727204081.76926: worker is 1 (out of 1 available) 13118 1727204081.76938: exiting _queue_task() for managed-node2/debug 13118 1727204081.76950: done queuing things up, now waiting for results queue to drain 13118 1727204081.76951: waiting for pending results... 13118 1727204081.77258: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections 13118 1727204081.77428: in run() - task 0affcd87-79f5-56a3-0a64-000000000039 13118 1727204081.77451: variable 'ansible_search_path' from source: unknown 13118 1727204081.77460: variable 'ansible_search_path' from source: unknown 13118 1727204081.77522: calling self._execute() 13118 1727204081.77629: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204081.77646: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204081.77659: variable 'omit' from source: magic vars 13118 1727204081.78073: variable 'ansible_distribution_major_version' from source: facts 13118 1727204081.78095: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204081.78106: variable 'omit' from source: magic vars 13118 1727204081.78172: variable 'omit' from source: magic vars 13118 1727204081.78218: variable 'omit' from source: magic vars 13118 1727204081.78265: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204081.78312: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204081.78338: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204081.78358: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204081.78377: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204081.78418: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204081.78426: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204081.78433: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204081.78543: Set connection var ansible_timeout to 10 13118 1727204081.78562: Set connection var ansible_pipelining to False 13118 1727204081.78573: Set connection var ansible_connection to ssh 13118 1727204081.78583: Set connection var ansible_shell_executable to /bin/sh 13118 1727204081.78592: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204081.78602: Set connection var ansible_shell_type to sh 13118 1727204081.78632: variable 'ansible_shell_executable' from source: unknown 13118 1727204081.78640: variable 'ansible_connection' from source: unknown 13118 1727204081.78646: variable 'ansible_module_compression' from source: unknown 13118 1727204081.78652: variable 'ansible_shell_type' from source: unknown 13118 1727204081.78658: variable 'ansible_shell_executable' from source: unknown 13118 1727204081.78665: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204081.78673: variable 'ansible_pipelining' from source: unknown 13118 1727204081.78679: variable 'ansible_timeout' from source: unknown 13118 1727204081.78686: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204081.78885: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204081.78903: variable 'omit' from source: magic vars 13118 1727204081.78934: starting attempt loop 13118 1727204081.78943: running the handler 13118 1727204081.79004: variable '__network_connections_result' from source: set_fact 13118 1727204081.79096: variable '__network_connections_result' from source: set_fact 13118 1727204081.79270: handler run complete 13118 1727204081.79300: attempt loop complete, returning result 13118 1727204081.79306: _execute() done 13118 1727204081.79311: dumping result to json 13118 1727204081.79319: done dumping result, returning 13118 1727204081.79328: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections [0affcd87-79f5-56a3-0a64-000000000039] 13118 1727204081.79337: sending task result for task 0affcd87-79f5-56a3-0a64-000000000039 13118 1727204081.79457: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000039 13118 1727204081.79465: WORKER PROCESS EXITING ok: [managed-node2] => { "__network_connections_result": { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "bond": { "miimon": 110, "mode": "active-backup" }, "interface_name": "nm-bond", "ip": { "route_metric4": 65535 }, "name": "bond0", "state": "up", "type": "bond" }, { "controller": "bond0", "interface_name": "test1", "name": "bond0.0", "state": "up", "type": "ethernet" }, { "controller": "bond0", "interface_name": "test2", "name": "bond0.1", "state": "up", "type": "ethernet" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true, "failed": false, "stderr": "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, debabe6d-f673-439e-9f14-25954ecc55a3\n[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 382c36b9-9f8f-4077-be2c-b77f342639e8\n[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 161d3351-eab4-440b-a3b9-ee4281b1f747\n[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, debabe6d-f673-439e-9f14-25954ecc55a3 (is-modified)\n[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 382c36b9-9f8f-4077-be2c-b77f342639e8 (not-active)\n[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 161d3351-eab4-440b-a3b9-ee4281b1f747 (not-active)\n", "stderr_lines": [ "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, debabe6d-f673-439e-9f14-25954ecc55a3", "[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 382c36b9-9f8f-4077-be2c-b77f342639e8", "[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 161d3351-eab4-440b-a3b9-ee4281b1f747", "[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, debabe6d-f673-439e-9f14-25954ecc55a3 (is-modified)", "[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 382c36b9-9f8f-4077-be2c-b77f342639e8 (not-active)", "[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 161d3351-eab4-440b-a3b9-ee4281b1f747 (not-active)" ] } } 13118 1727204081.79577: no more pending results, returning what we have 13118 1727204081.79582: results queue empty 13118 1727204081.79589: checking for any_errors_fatal 13118 1727204081.79594: done checking for any_errors_fatal 13118 1727204081.79595: checking for max_fail_percentage 13118 1727204081.79597: done checking for max_fail_percentage 13118 1727204081.79598: checking to see if all hosts have failed and the running result is not ok 13118 1727204081.79598: done checking to see if all hosts have failed 13118 1727204081.79599: getting the remaining hosts for this loop 13118 1727204081.79600: done getting the remaining hosts for this loop 13118 1727204081.79605: getting the next task for host managed-node2 13118 1727204081.79613: done getting next task for host managed-node2 13118 1727204081.79617: ^ task is: TASK: fedora.linux_system_roles.network : Show debug messages for the network_state 13118 1727204081.79621: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=24, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204081.79631: getting variables 13118 1727204081.79633: in VariableManager get_vars() 13118 1727204081.79681: Calling all_inventory to load vars for managed-node2 13118 1727204081.79684: Calling groups_inventory to load vars for managed-node2 13118 1727204081.79686: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204081.79697: Calling all_plugins_play to load vars for managed-node2 13118 1727204081.79700: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204081.79703: Calling groups_plugins_play to load vars for managed-node2 13118 1727204081.81699: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204081.85303: done with get_vars() 13118 1727204081.85338: done getting variables 13118 1727204081.85404: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show debug messages for the network_state] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:186 Tuesday 24 September 2024 14:54:41 -0400 (0:00:00.088) 0:00:19.154 ***** 13118 1727204081.85438: entering _queue_task() for managed-node2/debug 13118 1727204081.86283: worker is 1 (out of 1 available) 13118 1727204081.86294: exiting _queue_task() for managed-node2/debug 13118 1727204081.86306: done queuing things up, now waiting for results queue to drain 13118 1727204081.86307: waiting for pending results... 13118 1727204081.87959: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Show debug messages for the network_state 13118 1727204081.88106: in run() - task 0affcd87-79f5-56a3-0a64-00000000003a 13118 1727204081.88126: variable 'ansible_search_path' from source: unknown 13118 1727204081.88136: variable 'ansible_search_path' from source: unknown 13118 1727204081.88179: calling self._execute() 13118 1727204081.88270: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204081.88281: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204081.88294: variable 'omit' from source: magic vars 13118 1727204081.88684: variable 'ansible_distribution_major_version' from source: facts 13118 1727204081.88700: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204081.88834: variable 'network_state' from source: role '' defaults 13118 1727204081.88850: Evaluated conditional (network_state != {}): False 13118 1727204081.88858: when evaluation is False, skipping this task 13118 1727204081.88866: _execute() done 13118 1727204081.88873: dumping result to json 13118 1727204081.88879: done dumping result, returning 13118 1727204081.88889: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Show debug messages for the network_state [0affcd87-79f5-56a3-0a64-00000000003a] 13118 1727204081.88900: sending task result for task 0affcd87-79f5-56a3-0a64-00000000003a skipping: [managed-node2] => { "false_condition": "network_state != {}" } 13118 1727204081.89056: no more pending results, returning what we have 13118 1727204081.89061: results queue empty 13118 1727204081.89062: checking for any_errors_fatal 13118 1727204081.89075: done checking for any_errors_fatal 13118 1727204081.89075: checking for max_fail_percentage 13118 1727204081.89078: done checking for max_fail_percentage 13118 1727204081.89079: checking to see if all hosts have failed and the running result is not ok 13118 1727204081.89079: done checking to see if all hosts have failed 13118 1727204081.89080: getting the remaining hosts for this loop 13118 1727204081.89082: done getting the remaining hosts for this loop 13118 1727204081.89086: getting the next task for host managed-node2 13118 1727204081.89094: done getting next task for host managed-node2 13118 1727204081.89098: ^ task is: TASK: fedora.linux_system_roles.network : Re-test connectivity 13118 1727204081.89102: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=25, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204081.89119: getting variables 13118 1727204081.89121: in VariableManager get_vars() 13118 1727204081.89171: Calling all_inventory to load vars for managed-node2 13118 1727204081.89174: Calling groups_inventory to load vars for managed-node2 13118 1727204081.89177: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204081.89189: Calling all_plugins_play to load vars for managed-node2 13118 1727204081.89192: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204081.89195: Calling groups_plugins_play to load vars for managed-node2 13118 1727204081.90183: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000003a 13118 1727204081.90187: WORKER PROCESS EXITING 13118 1727204081.90907: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204081.92715: done with get_vars() 13118 1727204081.92742: done getting variables TASK [fedora.linux_system_roles.network : Re-test connectivity] **************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:192 Tuesday 24 September 2024 14:54:41 -0400 (0:00:00.074) 0:00:19.228 ***** 13118 1727204081.92847: entering _queue_task() for managed-node2/ping 13118 1727204081.92849: Creating lock for ping 13118 1727204081.93176: worker is 1 (out of 1 available) 13118 1727204081.93187: exiting _queue_task() for managed-node2/ping 13118 1727204081.93199: done queuing things up, now waiting for results queue to drain 13118 1727204081.93200: waiting for pending results... 13118 1727204081.93468: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Re-test connectivity 13118 1727204081.93596: in run() - task 0affcd87-79f5-56a3-0a64-00000000003b 13118 1727204081.93615: variable 'ansible_search_path' from source: unknown 13118 1727204081.93623: variable 'ansible_search_path' from source: unknown 13118 1727204081.93676: calling self._execute() 13118 1727204081.93777: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204081.93789: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204081.93801: variable 'omit' from source: magic vars 13118 1727204081.94407: variable 'ansible_distribution_major_version' from source: facts 13118 1727204081.94426: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204081.94440: variable 'omit' from source: magic vars 13118 1727204081.94502: variable 'omit' from source: magic vars 13118 1727204081.94731: variable 'omit' from source: magic vars 13118 1727204081.94780: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204081.94824: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204081.94858: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204081.94971: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204081.94987: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204081.95021: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204081.95062: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204081.95073: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204081.95298: Set connection var ansible_timeout to 10 13118 1727204081.95366: Set connection var ansible_pipelining to False 13118 1727204081.95374: Set connection var ansible_connection to ssh 13118 1727204081.95387: Set connection var ansible_shell_executable to /bin/sh 13118 1727204081.95396: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204081.95402: Set connection var ansible_shell_type to sh 13118 1727204081.95448: variable 'ansible_shell_executable' from source: unknown 13118 1727204081.95496: variable 'ansible_connection' from source: unknown 13118 1727204081.95504: variable 'ansible_module_compression' from source: unknown 13118 1727204081.95605: variable 'ansible_shell_type' from source: unknown 13118 1727204081.95614: variable 'ansible_shell_executable' from source: unknown 13118 1727204081.95621: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204081.95628: variable 'ansible_pipelining' from source: unknown 13118 1727204081.95638: variable 'ansible_timeout' from source: unknown 13118 1727204081.95645: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204081.95968: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204081.96048: variable 'omit' from source: magic vars 13118 1727204081.96057: starting attempt loop 13118 1727204081.96148: running the handler 13118 1727204081.96167: _low_level_execute_command(): starting 13118 1727204081.96180: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204081.98468: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204081.98472: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204081.99402: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204081.99407: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204081.99410: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204081.99480: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204081.99483: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204081.99491: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204081.99543: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204082.01235: stdout chunk (state=3): >>>/root <<< 13118 1727204082.01381: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204082.01440: stderr chunk (state=3): >>><<< 13118 1727204082.01443: stdout chunk (state=3): >>><<< 13118 1727204082.01562: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204082.01567: _low_level_execute_command(): starting 13118 1727204082.01570: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204082.0146713-15012-197964484965389 `" && echo ansible-tmp-1727204082.0146713-15012-197964484965389="` echo /root/.ansible/tmp/ansible-tmp-1727204082.0146713-15012-197964484965389 `" ) && sleep 0' 13118 1727204082.03363: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.03369: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.03408: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204082.03411: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.03423: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204082.03427: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.03649: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204082.03783: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204082.03789: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204082.03837: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204082.05738: stdout chunk (state=3): >>>ansible-tmp-1727204082.0146713-15012-197964484965389=/root/.ansible/tmp/ansible-tmp-1727204082.0146713-15012-197964484965389 <<< 13118 1727204082.05851: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204082.05936: stderr chunk (state=3): >>><<< 13118 1727204082.05939: stdout chunk (state=3): >>><<< 13118 1727204082.06273: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204082.0146713-15012-197964484965389=/root/.ansible/tmp/ansible-tmp-1727204082.0146713-15012-197964484965389 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204082.06277: variable 'ansible_module_compression' from source: unknown 13118 1727204082.06279: ANSIBALLZ: Using lock for ping 13118 1727204082.06281: ANSIBALLZ: Acquiring lock 13118 1727204082.06283: ANSIBALLZ: Lock acquired: 140051940598144 13118 1727204082.06286: ANSIBALLZ: Creating module 13118 1727204082.24881: ANSIBALLZ: Writing module into payload 13118 1727204082.24947: ANSIBALLZ: Writing module 13118 1727204082.24973: ANSIBALLZ: Renaming module 13118 1727204082.24982: ANSIBALLZ: Done creating module 13118 1727204082.24998: variable 'ansible_facts' from source: unknown 13118 1727204082.25065: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204082.0146713-15012-197964484965389/AnsiballZ_ping.py 13118 1727204082.25219: Sending initial data 13118 1727204082.25222: Sent initial data (153 bytes) 13118 1727204082.26212: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204082.26224: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.26235: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.26248: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.26295: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.26303: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204082.26313: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.26328: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204082.26334: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204082.26341: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204082.26348: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.26357: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.26370: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.26381: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.26388: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204082.26397: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.26469: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204082.26494: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204082.26507: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204082.26591: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204082.28481: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204082.28525: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204082.28571: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpdafah3qb /root/.ansible/tmp/ansible-tmp-1727204082.0146713-15012-197964484965389/AnsiballZ_ping.py <<< 13118 1727204082.28607: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204082.29688: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204082.29778: stderr chunk (state=3): >>><<< 13118 1727204082.29781: stdout chunk (state=3): >>><<< 13118 1727204082.29804: done transferring module to remote 13118 1727204082.29817: _low_level_execute_command(): starting 13118 1727204082.29821: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204082.0146713-15012-197964484965389/ /root/.ansible/tmp/ansible-tmp-1727204082.0146713-15012-197964484965389/AnsiballZ_ping.py && sleep 0' 13118 1727204082.30839: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.30843: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.30887: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204082.30893: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration <<< 13118 1727204082.30899: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.30919: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.30925: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.31012: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204082.31017: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204082.31035: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204082.31103: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204082.32990: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204082.32994: stdout chunk (state=3): >>><<< 13118 1727204082.33003: stderr chunk (state=3): >>><<< 13118 1727204082.33026: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204082.33032: _low_level_execute_command(): starting 13118 1727204082.33035: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204082.0146713-15012-197964484965389/AnsiballZ_ping.py && sleep 0' 13118 1727204082.33745: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204082.33755: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.33777: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.33792: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.33834: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.33838: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204082.33849: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.33863: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204082.33872: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204082.33889: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204082.33897: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.33909: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.33917: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.33927: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.33933: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204082.33940: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.34021: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204082.34049: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204082.34052: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204082.34378: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204082.47491: stdout chunk (state=3): >>> {"ping": "pong", "invocation": {"module_args": {"data": "pong"}}} <<< 13118 1727204082.48692: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204082.48696: stdout chunk (state=3): >>><<< 13118 1727204082.48698: stderr chunk (state=3): >>><<< 13118 1727204082.48834: _low_level_execute_command() done: rc=0, stdout= {"ping": "pong", "invocation": {"module_args": {"data": "pong"}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204082.48839: done with _execute_module (ping, {'_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ping', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204082.0146713-15012-197964484965389/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204082.48841: _low_level_execute_command(): starting 13118 1727204082.48843: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204082.0146713-15012-197964484965389/ > /dev/null 2>&1 && sleep 0' 13118 1727204082.49468: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204082.49488: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.49503: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.49521: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.49569: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.49581: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204082.49600: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.49623: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204082.49637: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204082.49648: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204082.49663: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.49679: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.49695: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.49718: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.49733: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204082.49748: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.49832: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204082.49857: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204082.49878: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204082.49955: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204082.51979: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204082.52002: stdout chunk (state=3): >>><<< 13118 1727204082.52006: stderr chunk (state=3): >>><<< 13118 1727204082.52271: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204082.52275: handler run complete 13118 1727204082.52278: attempt loop complete, returning result 13118 1727204082.52280: _execute() done 13118 1727204082.52282: dumping result to json 13118 1727204082.52284: done dumping result, returning 13118 1727204082.52287: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Re-test connectivity [0affcd87-79f5-56a3-0a64-00000000003b] 13118 1727204082.52289: sending task result for task 0affcd87-79f5-56a3-0a64-00000000003b 13118 1727204082.52362: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000003b ok: [managed-node2] => { "changed": false, "ping": "pong" } 13118 1727204082.52423: no more pending results, returning what we have 13118 1727204082.52427: results queue empty 13118 1727204082.52428: checking for any_errors_fatal 13118 1727204082.52433: done checking for any_errors_fatal 13118 1727204082.52434: checking for max_fail_percentage 13118 1727204082.52435: done checking for max_fail_percentage 13118 1727204082.52436: checking to see if all hosts have failed and the running result is not ok 13118 1727204082.52437: done checking to see if all hosts have failed 13118 1727204082.52438: getting the remaining hosts for this loop 13118 1727204082.52439: done getting the remaining hosts for this loop 13118 1727204082.52443: getting the next task for host managed-node2 13118 1727204082.52454: done getting next task for host managed-node2 13118 1727204082.52456: ^ task is: TASK: meta (role_complete) 13118 1727204082.52458: ^ state is: HOST STATE: block=2, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204082.52470: WORKER PROCESS EXITING 13118 1727204082.52476: getting variables 13118 1727204082.52478: in VariableManager get_vars() 13118 1727204082.52518: Calling all_inventory to load vars for managed-node2 13118 1727204082.52521: Calling groups_inventory to load vars for managed-node2 13118 1727204082.52523: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204082.52535: Calling all_plugins_play to load vars for managed-node2 13118 1727204082.52537: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204082.52540: Calling groups_plugins_play to load vars for managed-node2 13118 1727204082.54021: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204082.55679: done with get_vars() 13118 1727204082.55710: done getting variables 13118 1727204082.55801: done queuing things up, now waiting for results queue to drain 13118 1727204082.55804: results queue empty 13118 1727204082.55804: checking for any_errors_fatal 13118 1727204082.55808: done checking for any_errors_fatal 13118 1727204082.55808: checking for max_fail_percentage 13118 1727204082.55810: done checking for max_fail_percentage 13118 1727204082.55810: checking to see if all hosts have failed and the running result is not ok 13118 1727204082.55811: done checking to see if all hosts have failed 13118 1727204082.55812: getting the remaining hosts for this loop 13118 1727204082.55813: done getting the remaining hosts for this loop 13118 1727204082.55816: getting the next task for host managed-node2 13118 1727204082.55821: done getting next task for host managed-node2 13118 1727204082.55823: ^ task is: TASK: Include the task 'get_interface_stat.yml' 13118 1727204082.55826: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204082.55828: getting variables 13118 1727204082.55832: in VariableManager get_vars() 13118 1727204082.55848: Calling all_inventory to load vars for managed-node2 13118 1727204082.55850: Calling groups_inventory to load vars for managed-node2 13118 1727204082.55853: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204082.55858: Calling all_plugins_play to load vars for managed-node2 13118 1727204082.55860: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204082.55865: Calling groups_plugins_play to load vars for managed-node2 13118 1727204082.57207: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204082.58851: done with get_vars() 13118 1727204082.58880: done getting variables TASK [Include the task 'get_interface_stat.yml'] ******************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:3 Tuesday 24 September 2024 14:54:42 -0400 (0:00:00.661) 0:00:19.889 ***** 13118 1727204082.58976: entering _queue_task() for managed-node2/include_tasks 13118 1727204082.59368: worker is 1 (out of 1 available) 13118 1727204082.59381: exiting _queue_task() for managed-node2/include_tasks 13118 1727204082.59394: done queuing things up, now waiting for results queue to drain 13118 1727204082.59395: waiting for pending results... 13118 1727204082.59688: running TaskExecutor() for managed-node2/TASK: Include the task 'get_interface_stat.yml' 13118 1727204082.59834: in run() - task 0affcd87-79f5-56a3-0a64-00000000006e 13118 1727204082.59856: variable 'ansible_search_path' from source: unknown 13118 1727204082.59866: variable 'ansible_search_path' from source: unknown 13118 1727204082.59906: calling self._execute() 13118 1727204082.60008: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204082.60022: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204082.60039: variable 'omit' from source: magic vars 13118 1727204082.60427: variable 'ansible_distribution_major_version' from source: facts 13118 1727204082.60448: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204082.60460: _execute() done 13118 1727204082.60472: dumping result to json 13118 1727204082.60480: done dumping result, returning 13118 1727204082.60496: done running TaskExecutor() for managed-node2/TASK: Include the task 'get_interface_stat.yml' [0affcd87-79f5-56a3-0a64-00000000006e] 13118 1727204082.60507: sending task result for task 0affcd87-79f5-56a3-0a64-00000000006e 13118 1727204082.60638: no more pending results, returning what we have 13118 1727204082.60644: in VariableManager get_vars() 13118 1727204082.60694: Calling all_inventory to load vars for managed-node2 13118 1727204082.60698: Calling groups_inventory to load vars for managed-node2 13118 1727204082.60700: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204082.60716: Calling all_plugins_play to load vars for managed-node2 13118 1727204082.60719: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204082.60723: Calling groups_plugins_play to load vars for managed-node2 13118 1727204082.61783: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000006e 13118 1727204082.61787: WORKER PROCESS EXITING 13118 1727204082.62438: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204082.64096: done with get_vars() 13118 1727204082.64123: variable 'ansible_search_path' from source: unknown 13118 1727204082.64124: variable 'ansible_search_path' from source: unknown 13118 1727204082.64169: we have included files to process 13118 1727204082.64170: generating all_blocks data 13118 1727204082.64172: done generating all_blocks data 13118 1727204082.64177: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13118 1727204082.64178: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13118 1727204082.64180: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13118 1727204082.64385: done processing included file 13118 1727204082.64388: iterating over new_blocks loaded from include file 13118 1727204082.64389: in VariableManager get_vars() 13118 1727204082.64412: done with get_vars() 13118 1727204082.64414: filtering new block on tags 13118 1727204082.64434: done filtering new block on tags 13118 1727204082.64436: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml for managed-node2 13118 1727204082.64441: extending task lists for all hosts with included blocks 13118 1727204082.64556: done extending task lists 13118 1727204082.64558: done processing included files 13118 1727204082.64559: results queue empty 13118 1727204082.64559: checking for any_errors_fatal 13118 1727204082.64561: done checking for any_errors_fatal 13118 1727204082.64562: checking for max_fail_percentage 13118 1727204082.64563: done checking for max_fail_percentage 13118 1727204082.64566: checking to see if all hosts have failed and the running result is not ok 13118 1727204082.64567: done checking to see if all hosts have failed 13118 1727204082.64567: getting the remaining hosts for this loop 13118 1727204082.64568: done getting the remaining hosts for this loop 13118 1727204082.64571: getting the next task for host managed-node2 13118 1727204082.64576: done getting next task for host managed-node2 13118 1727204082.64578: ^ task is: TASK: Get stat for interface {{ interface }} 13118 1727204082.64580: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204082.64583: getting variables 13118 1727204082.64584: in VariableManager get_vars() 13118 1727204082.64599: Calling all_inventory to load vars for managed-node2 13118 1727204082.64601: Calling groups_inventory to load vars for managed-node2 13118 1727204082.64603: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204082.64609: Calling all_plugins_play to load vars for managed-node2 13118 1727204082.64611: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204082.64614: Calling groups_plugins_play to load vars for managed-node2 13118 1727204082.65910: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204082.66896: done with get_vars() 13118 1727204082.66914: done getting variables 13118 1727204082.67045: variable 'interface' from source: task vars 13118 1727204082.67048: variable 'controller_device' from source: play vars 13118 1727204082.67093: variable 'controller_device' from source: play vars TASK [Get stat for interface nm-bond] ****************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml:3 Tuesday 24 September 2024 14:54:42 -0400 (0:00:00.081) 0:00:19.971 ***** 13118 1727204082.67118: entering _queue_task() for managed-node2/stat 13118 1727204082.67358: worker is 1 (out of 1 available) 13118 1727204082.67373: exiting _queue_task() for managed-node2/stat 13118 1727204082.67384: done queuing things up, now waiting for results queue to drain 13118 1727204082.67385: waiting for pending results... 13118 1727204082.67565: running TaskExecutor() for managed-node2/TASK: Get stat for interface nm-bond 13118 1727204082.67656: in run() - task 0affcd87-79f5-56a3-0a64-000000000241 13118 1727204082.67668: variable 'ansible_search_path' from source: unknown 13118 1727204082.67672: variable 'ansible_search_path' from source: unknown 13118 1727204082.67713: calling self._execute() 13118 1727204082.67815: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204082.67831: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204082.67848: variable 'omit' from source: magic vars 13118 1727204082.68223: variable 'ansible_distribution_major_version' from source: facts 13118 1727204082.68244: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204082.68257: variable 'omit' from source: magic vars 13118 1727204082.68320: variable 'omit' from source: magic vars 13118 1727204082.68427: variable 'interface' from source: task vars 13118 1727204082.68440: variable 'controller_device' from source: play vars 13118 1727204082.68514: variable 'controller_device' from source: play vars 13118 1727204082.68540: variable 'omit' from source: magic vars 13118 1727204082.68589: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204082.68637: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204082.68666: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204082.68690: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204082.68710: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204082.68748: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204082.68757: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204082.68767: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204082.68878: Set connection var ansible_timeout to 10 13118 1727204082.68896: Set connection var ansible_pipelining to False 13118 1727204082.68904: Set connection var ansible_connection to ssh 13118 1727204082.68915: Set connection var ansible_shell_executable to /bin/sh 13118 1727204082.68932: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204082.68941: Set connection var ansible_shell_type to sh 13118 1727204082.68971: variable 'ansible_shell_executable' from source: unknown 13118 1727204082.68979: variable 'ansible_connection' from source: unknown 13118 1727204082.68985: variable 'ansible_module_compression' from source: unknown 13118 1727204082.68990: variable 'ansible_shell_type' from source: unknown 13118 1727204082.68996: variable 'ansible_shell_executable' from source: unknown 13118 1727204082.69001: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204082.69007: variable 'ansible_pipelining' from source: unknown 13118 1727204082.69014: variable 'ansible_timeout' from source: unknown 13118 1727204082.69021: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204082.69238: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204082.69248: variable 'omit' from source: magic vars 13118 1727204082.69253: starting attempt loop 13118 1727204082.69260: running the handler 13118 1727204082.69276: _low_level_execute_command(): starting 13118 1727204082.69282: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204082.70038: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.70049: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.70094: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204082.70100: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.70120: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.70125: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204082.70142: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.70219: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204082.70240: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204082.70243: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204082.70326: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204082.72006: stdout chunk (state=3): >>>/root <<< 13118 1727204082.72175: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204082.72185: stderr chunk (state=3): >>><<< 13118 1727204082.72188: stdout chunk (state=3): >>><<< 13118 1727204082.72215: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204082.72228: _low_level_execute_command(): starting 13118 1727204082.72237: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204082.7221413-15036-263563575008130 `" && echo ansible-tmp-1727204082.7221413-15036-263563575008130="` echo /root/.ansible/tmp/ansible-tmp-1727204082.7221413-15036-263563575008130 `" ) && sleep 0' 13118 1727204082.72908: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204082.72917: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.72928: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.72949: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.72990: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.72997: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204082.73007: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.73023: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204082.73031: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204082.73042: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204082.73051: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.73060: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.73074: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.73082: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.73090: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204082.73099: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.73176: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204082.73195: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204082.73207: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204082.73283: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204082.75202: stdout chunk (state=3): >>>ansible-tmp-1727204082.7221413-15036-263563575008130=/root/.ansible/tmp/ansible-tmp-1727204082.7221413-15036-263563575008130 <<< 13118 1727204082.75385: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204082.75434: stderr chunk (state=3): >>><<< 13118 1727204082.75437: stdout chunk (state=3): >>><<< 13118 1727204082.75670: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204082.7221413-15036-263563575008130=/root/.ansible/tmp/ansible-tmp-1727204082.7221413-15036-263563575008130 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204082.75674: variable 'ansible_module_compression' from source: unknown 13118 1727204082.75676: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13118 1727204082.75678: variable 'ansible_facts' from source: unknown 13118 1727204082.75709: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204082.7221413-15036-263563575008130/AnsiballZ_stat.py 13118 1727204082.75876: Sending initial data 13118 1727204082.75879: Sent initial data (153 bytes) 13118 1727204082.76959: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204082.76980: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.77007: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.77026: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.77072: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.77086: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204082.77107: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.77130: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204082.77143: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204082.77155: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204082.77169: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.77184: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.77199: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.77218: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.77235: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204082.77249: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.77334: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204082.77361: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204082.77382: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204082.77463: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204082.79287: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204082.79351: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204082.79385: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmp8uoljox_ /root/.ansible/tmp/ansible-tmp-1727204082.7221413-15036-263563575008130/AnsiballZ_stat.py <<< 13118 1727204082.79400: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204082.80431: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204082.80733: stderr chunk (state=3): >>><<< 13118 1727204082.80736: stdout chunk (state=3): >>><<< 13118 1727204082.80739: done transferring module to remote 13118 1727204082.80741: _low_level_execute_command(): starting 13118 1727204082.80743: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204082.7221413-15036-263563575008130/ /root/.ansible/tmp/ansible-tmp-1727204082.7221413-15036-263563575008130/AnsiballZ_stat.py && sleep 0' 13118 1727204082.81372: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204082.81396: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.81413: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.81432: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.81477: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.81495: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204082.81514: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.81531: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204082.81543: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204082.81553: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204082.81566: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.81579: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.81595: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.81619: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.81632: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204082.81646: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.81734: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204082.81757: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204082.81775: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204082.81850: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204082.84034: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204082.84141: stderr chunk (state=3): >>><<< 13118 1727204082.84157: stdout chunk (state=3): >>><<< 13118 1727204082.84276: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204082.84280: _low_level_execute_command(): starting 13118 1727204082.84282: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204082.7221413-15036-263563575008130/AnsiballZ_stat.py && sleep 0' 13118 1727204082.84900: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204082.84918: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.84942: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.84960: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.85004: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.85016: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204082.85028: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.85054: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204082.85068: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204082.85079: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204082.85090: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204082.85102: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204082.85117: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204082.85128: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204082.85140: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204082.85160: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204082.85238: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204082.85272: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204082.85289: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204082.85379: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204082.98913: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/nm-bond", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 26340, "dev": 21, "nlink": 1, "atime": 1727204081.2865894, "mtime": 1727204081.2865894, "ctime": 1727204081.2865894, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/nm-bond", "lnk_target": "../../devices/virtual/net/nm-bond", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/nm-bond", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13118 1727204082.99987: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204083.00096: stderr chunk (state=3): >>><<< 13118 1727204083.00100: stdout chunk (state=3): >>><<< 13118 1727204083.00171: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/nm-bond", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 26340, "dev": 21, "nlink": 1, "atime": 1727204081.2865894, "mtime": 1727204081.2865894, "ctime": 1727204081.2865894, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/nm-bond", "lnk_target": "../../devices/virtual/net/nm-bond", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/nm-bond", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204083.00290: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/sys/class/net/nm-bond', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204082.7221413-15036-263563575008130/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204083.00294: _low_level_execute_command(): starting 13118 1727204083.00296: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204082.7221413-15036-263563575008130/ > /dev/null 2>&1 && sleep 0' 13118 1727204083.00969: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204083.00983: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.00997: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.01015: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.01108: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.01111: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.01116: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.01119: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.01185: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204083.01197: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204083.01247: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204083.03124: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204083.03187: stderr chunk (state=3): >>><<< 13118 1727204083.03191: stdout chunk (state=3): >>><<< 13118 1727204083.03209: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204083.03216: handler run complete 13118 1727204083.03269: attempt loop complete, returning result 13118 1727204083.03273: _execute() done 13118 1727204083.03277: dumping result to json 13118 1727204083.03283: done dumping result, returning 13118 1727204083.03291: done running TaskExecutor() for managed-node2/TASK: Get stat for interface nm-bond [0affcd87-79f5-56a3-0a64-000000000241] 13118 1727204083.03297: sending task result for task 0affcd87-79f5-56a3-0a64-000000000241 13118 1727204083.03413: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000241 13118 1727204083.03415: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "stat": { "atime": 1727204081.2865894, "block_size": 4096, "blocks": 0, "ctime": 1727204081.2865894, "dev": 21, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 26340, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": true, "isreg": false, "issock": false, "isuid": false, "lnk_source": "/sys/devices/virtual/net/nm-bond", "lnk_target": "../../devices/virtual/net/nm-bond", "mode": "0777", "mtime": 1727204081.2865894, "nlink": 1, "path": "/sys/class/net/nm-bond", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "wgrp": true, "woth": true, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } 13118 1727204083.03503: no more pending results, returning what we have 13118 1727204083.03508: results queue empty 13118 1727204083.03509: checking for any_errors_fatal 13118 1727204083.03510: done checking for any_errors_fatal 13118 1727204083.03511: checking for max_fail_percentage 13118 1727204083.03513: done checking for max_fail_percentage 13118 1727204083.03513: checking to see if all hosts have failed and the running result is not ok 13118 1727204083.03514: done checking to see if all hosts have failed 13118 1727204083.03515: getting the remaining hosts for this loop 13118 1727204083.03516: done getting the remaining hosts for this loop 13118 1727204083.03520: getting the next task for host managed-node2 13118 1727204083.03527: done getting next task for host managed-node2 13118 1727204083.03530: ^ task is: TASK: Assert that the interface is present - '{{ interface }}' 13118 1727204083.03533: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204083.03537: getting variables 13118 1727204083.03538: in VariableManager get_vars() 13118 1727204083.03578: Calling all_inventory to load vars for managed-node2 13118 1727204083.03581: Calling groups_inventory to load vars for managed-node2 13118 1727204083.03583: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204083.03593: Calling all_plugins_play to load vars for managed-node2 13118 1727204083.03596: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204083.03599: Calling groups_plugins_play to load vars for managed-node2 13118 1727204083.04694: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204083.05709: done with get_vars() 13118 1727204083.05727: done getting variables 13118 1727204083.05774: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204083.05863: variable 'interface' from source: task vars 13118 1727204083.05868: variable 'controller_device' from source: play vars 13118 1727204083.05910: variable 'controller_device' from source: play vars TASK [Assert that the interface is present - 'nm-bond'] ************************ task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:5 Tuesday 24 September 2024 14:54:43 -0400 (0:00:00.388) 0:00:20.359 ***** 13118 1727204083.05935: entering _queue_task() for managed-node2/assert 13118 1727204083.06170: worker is 1 (out of 1 available) 13118 1727204083.06184: exiting _queue_task() for managed-node2/assert 13118 1727204083.06196: done queuing things up, now waiting for results queue to drain 13118 1727204083.06198: waiting for pending results... 13118 1727204083.06386: running TaskExecutor() for managed-node2/TASK: Assert that the interface is present - 'nm-bond' 13118 1727204083.06471: in run() - task 0affcd87-79f5-56a3-0a64-00000000006f 13118 1727204083.06482: variable 'ansible_search_path' from source: unknown 13118 1727204083.06486: variable 'ansible_search_path' from source: unknown 13118 1727204083.06517: calling self._execute() 13118 1727204083.06584: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.06589: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.06597: variable 'omit' from source: magic vars 13118 1727204083.06866: variable 'ansible_distribution_major_version' from source: facts 13118 1727204083.06877: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204083.06882: variable 'omit' from source: magic vars 13118 1727204083.06914: variable 'omit' from source: magic vars 13118 1727204083.06987: variable 'interface' from source: task vars 13118 1727204083.06991: variable 'controller_device' from source: play vars 13118 1727204083.07039: variable 'controller_device' from source: play vars 13118 1727204083.07055: variable 'omit' from source: magic vars 13118 1727204083.07091: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204083.07117: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204083.07136: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204083.07153: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204083.07169: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204083.07192: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204083.07195: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.07197: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.07265: Set connection var ansible_timeout to 10 13118 1727204083.07278: Set connection var ansible_pipelining to False 13118 1727204083.07281: Set connection var ansible_connection to ssh 13118 1727204083.07285: Set connection var ansible_shell_executable to /bin/sh 13118 1727204083.07288: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204083.07290: Set connection var ansible_shell_type to sh 13118 1727204083.07308: variable 'ansible_shell_executable' from source: unknown 13118 1727204083.07311: variable 'ansible_connection' from source: unknown 13118 1727204083.07313: variable 'ansible_module_compression' from source: unknown 13118 1727204083.07316: variable 'ansible_shell_type' from source: unknown 13118 1727204083.07318: variable 'ansible_shell_executable' from source: unknown 13118 1727204083.07320: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.07325: variable 'ansible_pipelining' from source: unknown 13118 1727204083.07327: variable 'ansible_timeout' from source: unknown 13118 1727204083.07333: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.07433: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204083.07444: variable 'omit' from source: magic vars 13118 1727204083.07449: starting attempt loop 13118 1727204083.07452: running the handler 13118 1727204083.07546: variable 'interface_stat' from source: set_fact 13118 1727204083.07560: Evaluated conditional (interface_stat.stat.exists): True 13118 1727204083.07567: handler run complete 13118 1727204083.07578: attempt loop complete, returning result 13118 1727204083.07581: _execute() done 13118 1727204083.07583: dumping result to json 13118 1727204083.07586: done dumping result, returning 13118 1727204083.07593: done running TaskExecutor() for managed-node2/TASK: Assert that the interface is present - 'nm-bond' [0affcd87-79f5-56a3-0a64-00000000006f] 13118 1727204083.07598: sending task result for task 0affcd87-79f5-56a3-0a64-00000000006f 13118 1727204083.07687: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000006f 13118 1727204083.07689: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false } MSG: All assertions passed 13118 1727204083.07769: no more pending results, returning what we have 13118 1727204083.07773: results queue empty 13118 1727204083.07775: checking for any_errors_fatal 13118 1727204083.07782: done checking for any_errors_fatal 13118 1727204083.07783: checking for max_fail_percentage 13118 1727204083.07784: done checking for max_fail_percentage 13118 1727204083.07785: checking to see if all hosts have failed and the running result is not ok 13118 1727204083.07786: done checking to see if all hosts have failed 13118 1727204083.07787: getting the remaining hosts for this loop 13118 1727204083.07788: done getting the remaining hosts for this loop 13118 1727204083.07791: getting the next task for host managed-node2 13118 1727204083.07799: done getting next task for host managed-node2 13118 1727204083.07802: ^ task is: TASK: Include the task 'assert_profile_present.yml' 13118 1727204083.07803: ^ state is: HOST STATE: block=2, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204083.07807: getting variables 13118 1727204083.07808: in VariableManager get_vars() 13118 1727204083.07853: Calling all_inventory to load vars for managed-node2 13118 1727204083.07855: Calling groups_inventory to load vars for managed-node2 13118 1727204083.07858: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204083.07869: Calling all_plugins_play to load vars for managed-node2 13118 1727204083.07872: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204083.07874: Calling groups_plugins_play to load vars for managed-node2 13118 1727204083.08678: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204083.09599: done with get_vars() 13118 1727204083.09618: done getting variables TASK [Include the task 'assert_profile_present.yml'] *************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml:67 Tuesday 24 September 2024 14:54:43 -0400 (0:00:00.037) 0:00:20.396 ***** 13118 1727204083.09693: entering _queue_task() for managed-node2/include_tasks 13118 1727204083.09932: worker is 1 (out of 1 available) 13118 1727204083.09945: exiting _queue_task() for managed-node2/include_tasks 13118 1727204083.09957: done queuing things up, now waiting for results queue to drain 13118 1727204083.09959: waiting for pending results... 13118 1727204083.10140: running TaskExecutor() for managed-node2/TASK: Include the task 'assert_profile_present.yml' 13118 1727204083.10200: in run() - task 0affcd87-79f5-56a3-0a64-000000000070 13118 1727204083.10211: variable 'ansible_search_path' from source: unknown 13118 1727204083.10249: variable 'controller_profile' from source: play vars 13118 1727204083.10395: variable 'controller_profile' from source: play vars 13118 1727204083.10406: variable 'port1_profile' from source: play vars 13118 1727204083.10464: variable 'port1_profile' from source: play vars 13118 1727204083.10473: variable 'port2_profile' from source: play vars 13118 1727204083.10523: variable 'port2_profile' from source: play vars 13118 1727204083.10540: variable 'omit' from source: magic vars 13118 1727204083.10644: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.10648: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.10657: variable 'omit' from source: magic vars 13118 1727204083.10822: variable 'ansible_distribution_major_version' from source: facts 13118 1727204083.10830: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204083.10855: variable 'item' from source: unknown 13118 1727204083.10902: variable 'item' from source: unknown 13118 1727204083.11020: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.11023: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.11026: variable 'omit' from source: magic vars 13118 1727204083.11109: variable 'ansible_distribution_major_version' from source: facts 13118 1727204083.11112: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204083.11131: variable 'item' from source: unknown 13118 1727204083.11177: variable 'item' from source: unknown 13118 1727204083.11245: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.11250: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.11259: variable 'omit' from source: magic vars 13118 1727204083.11363: variable 'ansible_distribution_major_version' from source: facts 13118 1727204083.11366: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204083.11387: variable 'item' from source: unknown 13118 1727204083.11434: variable 'item' from source: unknown 13118 1727204083.11497: dumping result to json 13118 1727204083.11500: done dumping result, returning 13118 1727204083.11503: done running TaskExecutor() for managed-node2/TASK: Include the task 'assert_profile_present.yml' [0affcd87-79f5-56a3-0a64-000000000070] 13118 1727204083.11505: sending task result for task 0affcd87-79f5-56a3-0a64-000000000070 13118 1727204083.11548: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000070 13118 1727204083.11550: WORKER PROCESS EXITING 13118 1727204083.11579: no more pending results, returning what we have 13118 1727204083.11584: in VariableManager get_vars() 13118 1727204083.11638: Calling all_inventory to load vars for managed-node2 13118 1727204083.11641: Calling groups_inventory to load vars for managed-node2 13118 1727204083.11643: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204083.11656: Calling all_plugins_play to load vars for managed-node2 13118 1727204083.11659: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204083.11661: Calling groups_plugins_play to load vars for managed-node2 13118 1727204083.12616: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204083.13519: done with get_vars() 13118 1727204083.13540: variable 'ansible_search_path' from source: unknown 13118 1727204083.13554: variable 'ansible_search_path' from source: unknown 13118 1727204083.13559: variable 'ansible_search_path' from source: unknown 13118 1727204083.13566: we have included files to process 13118 1727204083.13567: generating all_blocks data 13118 1727204083.13568: done generating all_blocks data 13118 1727204083.13571: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13118 1727204083.13571: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13118 1727204083.13573: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13118 1727204083.13714: in VariableManager get_vars() 13118 1727204083.13734: done with get_vars() 13118 1727204083.13910: done processing included file 13118 1727204083.13912: iterating over new_blocks loaded from include file 13118 1727204083.13913: in VariableManager get_vars() 13118 1727204083.13925: done with get_vars() 13118 1727204083.13926: filtering new block on tags 13118 1727204083.13941: done filtering new block on tags 13118 1727204083.13943: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml for managed-node2 => (item=bond0) 13118 1727204083.13946: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13118 1727204083.13947: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13118 1727204083.13949: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13118 1727204083.14011: in VariableManager get_vars() 13118 1727204083.14026: done with get_vars() 13118 1727204083.14179: done processing included file 13118 1727204083.14180: iterating over new_blocks loaded from include file 13118 1727204083.14181: in VariableManager get_vars() 13118 1727204083.14192: done with get_vars() 13118 1727204083.14193: filtering new block on tags 13118 1727204083.14204: done filtering new block on tags 13118 1727204083.14205: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml for managed-node2 => (item=bond0.0) 13118 1727204083.14208: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13118 1727204083.14209: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13118 1727204083.14211: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13118 1727204083.14277: in VariableManager get_vars() 13118 1727204083.14334: done with get_vars() 13118 1727204083.14486: done processing included file 13118 1727204083.14487: iterating over new_blocks loaded from include file 13118 1727204083.14488: in VariableManager get_vars() 13118 1727204083.14499: done with get_vars() 13118 1727204083.14500: filtering new block on tags 13118 1727204083.14511: done filtering new block on tags 13118 1727204083.14512: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml for managed-node2 => (item=bond0.1) 13118 1727204083.14515: extending task lists for all hosts with included blocks 13118 1727204083.16115: done extending task lists 13118 1727204083.16122: done processing included files 13118 1727204083.16123: results queue empty 13118 1727204083.16123: checking for any_errors_fatal 13118 1727204083.16126: done checking for any_errors_fatal 13118 1727204083.16127: checking for max_fail_percentage 13118 1727204083.16127: done checking for max_fail_percentage 13118 1727204083.16128: checking to see if all hosts have failed and the running result is not ok 13118 1727204083.16128: done checking to see if all hosts have failed 13118 1727204083.16131: getting the remaining hosts for this loop 13118 1727204083.16132: done getting the remaining hosts for this loop 13118 1727204083.16134: getting the next task for host managed-node2 13118 1727204083.16137: done getting next task for host managed-node2 13118 1727204083.16138: ^ task is: TASK: Include the task 'get_profile_stat.yml' 13118 1727204083.16140: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204083.16142: getting variables 13118 1727204083.16143: in VariableManager get_vars() 13118 1727204083.16156: Calling all_inventory to load vars for managed-node2 13118 1727204083.16157: Calling groups_inventory to load vars for managed-node2 13118 1727204083.16158: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204083.16166: Calling all_plugins_play to load vars for managed-node2 13118 1727204083.16167: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204083.16169: Calling groups_plugins_play to load vars for managed-node2 13118 1727204083.16873: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204083.17861: done with get_vars() 13118 1727204083.17878: done getting variables TASK [Include the task 'get_profile_stat.yml'] ********************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:3 Tuesday 24 September 2024 14:54:43 -0400 (0:00:00.082) 0:00:20.479 ***** 13118 1727204083.17941: entering _queue_task() for managed-node2/include_tasks 13118 1727204083.18192: worker is 1 (out of 1 available) 13118 1727204083.18206: exiting _queue_task() for managed-node2/include_tasks 13118 1727204083.18218: done queuing things up, now waiting for results queue to drain 13118 1727204083.18219: waiting for pending results... 13118 1727204083.18404: running TaskExecutor() for managed-node2/TASK: Include the task 'get_profile_stat.yml' 13118 1727204083.18472: in run() - task 0affcd87-79f5-56a3-0a64-00000000025f 13118 1727204083.18483: variable 'ansible_search_path' from source: unknown 13118 1727204083.18486: variable 'ansible_search_path' from source: unknown 13118 1727204083.18515: calling self._execute() 13118 1727204083.18584: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.18588: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.18597: variable 'omit' from source: magic vars 13118 1727204083.18866: variable 'ansible_distribution_major_version' from source: facts 13118 1727204083.18877: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204083.18880: _execute() done 13118 1727204083.18885: dumping result to json 13118 1727204083.18896: done dumping result, returning 13118 1727204083.18901: done running TaskExecutor() for managed-node2/TASK: Include the task 'get_profile_stat.yml' [0affcd87-79f5-56a3-0a64-00000000025f] 13118 1727204083.18908: sending task result for task 0affcd87-79f5-56a3-0a64-00000000025f 13118 1727204083.18991: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000025f 13118 1727204083.18995: WORKER PROCESS EXITING 13118 1727204083.19020: no more pending results, returning what we have 13118 1727204083.19026: in VariableManager get_vars() 13118 1727204083.19072: Calling all_inventory to load vars for managed-node2 13118 1727204083.19075: Calling groups_inventory to load vars for managed-node2 13118 1727204083.19077: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204083.19090: Calling all_plugins_play to load vars for managed-node2 13118 1727204083.19093: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204083.19095: Calling groups_plugins_play to load vars for managed-node2 13118 1727204083.23073: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204083.23963: done with get_vars() 13118 1727204083.23980: variable 'ansible_search_path' from source: unknown 13118 1727204083.23981: variable 'ansible_search_path' from source: unknown 13118 1727204083.24009: we have included files to process 13118 1727204083.24010: generating all_blocks data 13118 1727204083.24010: done generating all_blocks data 13118 1727204083.24011: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13118 1727204083.24012: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13118 1727204083.24013: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13118 1727204083.24667: done processing included file 13118 1727204083.24669: iterating over new_blocks loaded from include file 13118 1727204083.24670: in VariableManager get_vars() 13118 1727204083.24685: done with get_vars() 13118 1727204083.24686: filtering new block on tags 13118 1727204083.24700: done filtering new block on tags 13118 1727204083.24702: in VariableManager get_vars() 13118 1727204083.24712: done with get_vars() 13118 1727204083.24713: filtering new block on tags 13118 1727204083.24727: done filtering new block on tags 13118 1727204083.24728: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml for managed-node2 13118 1727204083.24731: extending task lists for all hosts with included blocks 13118 1727204083.24877: done extending task lists 13118 1727204083.24878: done processing included files 13118 1727204083.24879: results queue empty 13118 1727204083.24879: checking for any_errors_fatal 13118 1727204083.24881: done checking for any_errors_fatal 13118 1727204083.24882: checking for max_fail_percentage 13118 1727204083.24882: done checking for max_fail_percentage 13118 1727204083.24883: checking to see if all hosts have failed and the running result is not ok 13118 1727204083.24883: done checking to see if all hosts have failed 13118 1727204083.24884: getting the remaining hosts for this loop 13118 1727204083.24885: done getting the remaining hosts for this loop 13118 1727204083.24886: getting the next task for host managed-node2 13118 1727204083.24888: done getting next task for host managed-node2 13118 1727204083.24890: ^ task is: TASK: Initialize NM profile exist and ansible_managed comment flag 13118 1727204083.24892: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204083.24893: getting variables 13118 1727204083.24893: in VariableManager get_vars() 13118 1727204083.24903: Calling all_inventory to load vars for managed-node2 13118 1727204083.24904: Calling groups_inventory to load vars for managed-node2 13118 1727204083.24906: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204083.24910: Calling all_plugins_play to load vars for managed-node2 13118 1727204083.24911: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204083.24913: Calling groups_plugins_play to load vars for managed-node2 13118 1727204083.25589: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204083.26486: done with get_vars() 13118 1727204083.26501: done getting variables 13118 1727204083.26528: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Initialize NM profile exist and ansible_managed comment flag] ************ task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:3 Tuesday 24 September 2024 14:54:43 -0400 (0:00:00.086) 0:00:20.565 ***** 13118 1727204083.26547: entering _queue_task() for managed-node2/set_fact 13118 1727204083.26786: worker is 1 (out of 1 available) 13118 1727204083.26798: exiting _queue_task() for managed-node2/set_fact 13118 1727204083.26810: done queuing things up, now waiting for results queue to drain 13118 1727204083.26812: waiting for pending results... 13118 1727204083.27004: running TaskExecutor() for managed-node2/TASK: Initialize NM profile exist and ansible_managed comment flag 13118 1727204083.27074: in run() - task 0affcd87-79f5-56a3-0a64-0000000003b0 13118 1727204083.27084: variable 'ansible_search_path' from source: unknown 13118 1727204083.27088: variable 'ansible_search_path' from source: unknown 13118 1727204083.27115: calling self._execute() 13118 1727204083.27189: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.27193: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.27201: variable 'omit' from source: magic vars 13118 1727204083.27479: variable 'ansible_distribution_major_version' from source: facts 13118 1727204083.27489: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204083.27494: variable 'omit' from source: magic vars 13118 1727204083.27527: variable 'omit' from source: magic vars 13118 1727204083.27553: variable 'omit' from source: magic vars 13118 1727204083.27590: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204083.27619: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204083.27636: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204083.27651: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204083.27661: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204083.27689: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204083.27693: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.27695: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.27763: Set connection var ansible_timeout to 10 13118 1727204083.27773: Set connection var ansible_pipelining to False 13118 1727204083.27778: Set connection var ansible_connection to ssh 13118 1727204083.27781: Set connection var ansible_shell_executable to /bin/sh 13118 1727204083.27785: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204083.27794: Set connection var ansible_shell_type to sh 13118 1727204083.27811: variable 'ansible_shell_executable' from source: unknown 13118 1727204083.27815: variable 'ansible_connection' from source: unknown 13118 1727204083.27817: variable 'ansible_module_compression' from source: unknown 13118 1727204083.27820: variable 'ansible_shell_type' from source: unknown 13118 1727204083.27823: variable 'ansible_shell_executable' from source: unknown 13118 1727204083.27825: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.27828: variable 'ansible_pipelining' from source: unknown 13118 1727204083.27834: variable 'ansible_timeout' from source: unknown 13118 1727204083.27836: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.27940: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204083.27949: variable 'omit' from source: magic vars 13118 1727204083.27955: starting attempt loop 13118 1727204083.27958: running the handler 13118 1727204083.27972: handler run complete 13118 1727204083.27980: attempt loop complete, returning result 13118 1727204083.27983: _execute() done 13118 1727204083.27986: dumping result to json 13118 1727204083.27988: done dumping result, returning 13118 1727204083.27994: done running TaskExecutor() for managed-node2/TASK: Initialize NM profile exist and ansible_managed comment flag [0affcd87-79f5-56a3-0a64-0000000003b0] 13118 1727204083.28001: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b0 13118 1727204083.28083: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b0 13118 1727204083.28086: WORKER PROCESS EXITING ok: [managed-node2] => { "ansible_facts": { "lsr_net_profile_ansible_managed": false, "lsr_net_profile_exists": false, "lsr_net_profile_fingerprint": false }, "changed": false } 13118 1727204083.28145: no more pending results, returning what we have 13118 1727204083.28149: results queue empty 13118 1727204083.28150: checking for any_errors_fatal 13118 1727204083.28152: done checking for any_errors_fatal 13118 1727204083.28153: checking for max_fail_percentage 13118 1727204083.28154: done checking for max_fail_percentage 13118 1727204083.28155: checking to see if all hosts have failed and the running result is not ok 13118 1727204083.28156: done checking to see if all hosts have failed 13118 1727204083.28157: getting the remaining hosts for this loop 13118 1727204083.28158: done getting the remaining hosts for this loop 13118 1727204083.28161: getting the next task for host managed-node2 13118 1727204083.28171: done getting next task for host managed-node2 13118 1727204083.28173: ^ task is: TASK: Stat profile file 13118 1727204083.28178: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204083.28182: getting variables 13118 1727204083.28183: in VariableManager get_vars() 13118 1727204083.28228: Calling all_inventory to load vars for managed-node2 13118 1727204083.28233: Calling groups_inventory to load vars for managed-node2 13118 1727204083.28235: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204083.28245: Calling all_plugins_play to load vars for managed-node2 13118 1727204083.28247: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204083.28250: Calling groups_plugins_play to load vars for managed-node2 13118 1727204083.29108: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204083.30036: done with get_vars() 13118 1727204083.30054: done getting variables TASK [Stat profile file] ******************************************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:9 Tuesday 24 September 2024 14:54:43 -0400 (0:00:00.035) 0:00:20.601 ***** 13118 1727204083.30125: entering _queue_task() for managed-node2/stat 13118 1727204083.30356: worker is 1 (out of 1 available) 13118 1727204083.30369: exiting _queue_task() for managed-node2/stat 13118 1727204083.30382: done queuing things up, now waiting for results queue to drain 13118 1727204083.30383: waiting for pending results... 13118 1727204083.30603: running TaskExecutor() for managed-node2/TASK: Stat profile file 13118 1727204083.30725: in run() - task 0affcd87-79f5-56a3-0a64-0000000003b1 13118 1727204083.30752: variable 'ansible_search_path' from source: unknown 13118 1727204083.30759: variable 'ansible_search_path' from source: unknown 13118 1727204083.30797: calling self._execute() 13118 1727204083.30899: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.30910: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.30924: variable 'omit' from source: magic vars 13118 1727204083.31316: variable 'ansible_distribution_major_version' from source: facts 13118 1727204083.31335: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204083.31346: variable 'omit' from source: magic vars 13118 1727204083.31403: variable 'omit' from source: magic vars 13118 1727204083.31506: variable 'profile' from source: include params 13118 1727204083.31515: variable 'item' from source: include params 13118 1727204083.31587: variable 'item' from source: include params 13118 1727204083.31618: variable 'omit' from source: magic vars 13118 1727204083.31666: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204083.31710: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204083.31740: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204083.31763: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204083.31782: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204083.31818: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204083.31830: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.31838: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.31949: Set connection var ansible_timeout to 10 13118 1727204083.31968: Set connection var ansible_pipelining to False 13118 1727204083.31975: Set connection var ansible_connection to ssh 13118 1727204083.31986: Set connection var ansible_shell_executable to /bin/sh 13118 1727204083.31995: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204083.32002: Set connection var ansible_shell_type to sh 13118 1727204083.32047: variable 'ansible_shell_executable' from source: unknown 13118 1727204083.32058: variable 'ansible_connection' from source: unknown 13118 1727204083.32079: variable 'ansible_module_compression' from source: unknown 13118 1727204083.32083: variable 'ansible_shell_type' from source: unknown 13118 1727204083.32086: variable 'ansible_shell_executable' from source: unknown 13118 1727204083.32088: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.32095: variable 'ansible_pipelining' from source: unknown 13118 1727204083.32116: variable 'ansible_timeout' from source: unknown 13118 1727204083.32119: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.32274: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204083.32283: variable 'omit' from source: magic vars 13118 1727204083.32289: starting attempt loop 13118 1727204083.32292: running the handler 13118 1727204083.32306: _low_level_execute_command(): starting 13118 1727204083.32311: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204083.32826: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.32844: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.32862: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.32878: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.32928: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204083.32944: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204083.32997: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204083.34665: stdout chunk (state=3): >>>/root <<< 13118 1727204083.34848: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204083.34861: stdout chunk (state=3): >>><<< 13118 1727204083.34876: stderr chunk (state=3): >>><<< 13118 1727204083.34902: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204083.34921: _low_level_execute_command(): starting 13118 1727204083.34934: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204083.3490949-15060-59900587035213 `" && echo ansible-tmp-1727204083.3490949-15060-59900587035213="` echo /root/.ansible/tmp/ansible-tmp-1727204083.3490949-15060-59900587035213 `" ) && sleep 0' 13118 1727204083.35587: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204083.35601: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.35617: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.35637: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.35685: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.35708: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204083.35726: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.35743: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204083.35754: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204083.35761: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204083.35774: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.35784: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.35796: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.35805: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.35813: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204083.35823: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.35899: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204083.35917: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204083.35929: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204083.36002: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204083.37892: stdout chunk (state=3): >>>ansible-tmp-1727204083.3490949-15060-59900587035213=/root/.ansible/tmp/ansible-tmp-1727204083.3490949-15060-59900587035213 <<< 13118 1727204083.38083: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204083.38087: stdout chunk (state=3): >>><<< 13118 1727204083.38092: stderr chunk (state=3): >>><<< 13118 1727204083.38271: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204083.3490949-15060-59900587035213=/root/.ansible/tmp/ansible-tmp-1727204083.3490949-15060-59900587035213 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204083.38275: variable 'ansible_module_compression' from source: unknown 13118 1727204083.38278: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13118 1727204083.38394: variable 'ansible_facts' from source: unknown 13118 1727204083.38407: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204083.3490949-15060-59900587035213/AnsiballZ_stat.py 13118 1727204083.38589: Sending initial data 13118 1727204083.38592: Sent initial data (152 bytes) 13118 1727204083.39674: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204083.39688: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.39708: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.39727: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.39771: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.39782: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204083.39794: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.39817: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204083.39826: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204083.39838: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204083.39847: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.39858: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.39875: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.39885: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.39893: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204083.39903: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.39990: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204083.40006: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204083.40024: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204083.40095: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204083.41893: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204083.41973: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204083.41977: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmp3hddfrwk /root/.ansible/tmp/ansible-tmp-1727204083.3490949-15060-59900587035213/AnsiballZ_stat.py <<< 13118 1727204083.42312: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204083.43370: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204083.43641: stderr chunk (state=3): >>><<< 13118 1727204083.43645: stdout chunk (state=3): >>><<< 13118 1727204083.43647: done transferring module to remote 13118 1727204083.43649: _low_level_execute_command(): starting 13118 1727204083.43651: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204083.3490949-15060-59900587035213/ /root/.ansible/tmp/ansible-tmp-1727204083.3490949-15060-59900587035213/AnsiballZ_stat.py && sleep 0' 13118 1727204083.44247: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204083.44263: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.44280: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.44302: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.44348: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.44360: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204083.44376: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.44394: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204083.44405: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204083.44416: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204083.44433: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.44447: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.44467: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.44479: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.44489: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204083.44501: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.44586: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204083.44605: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204083.44619: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204083.44690: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204083.46462: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204083.46552: stderr chunk (state=3): >>><<< 13118 1727204083.46555: stdout chunk (state=3): >>><<< 13118 1727204083.46656: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204083.46659: _low_level_execute_command(): starting 13118 1727204083.46662: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204083.3490949-15060-59900587035213/AnsiballZ_stat.py && sleep 0' 13118 1727204083.47266: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204083.47282: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.47298: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.47317: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.47368: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.47382: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204083.47397: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.47415: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204083.47428: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204083.47443: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204083.47455: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.47470: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.47489: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.47501: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.47511: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204083.47524: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.47614: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204083.47634: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204083.47649: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204083.47746: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204083.61109: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/etc/sysconfig/network-scripts/ifcfg-bond0", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13118 1727204083.62235: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204083.62239: stdout chunk (state=3): >>><<< 13118 1727204083.62242: stderr chunk (state=3): >>><<< 13118 1727204083.62273: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/etc/sysconfig/network-scripts/ifcfg-bond0", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204083.62393: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/etc/sysconfig/network-scripts/ifcfg-bond0', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204083.3490949-15060-59900587035213/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204083.62401: _low_level_execute_command(): starting 13118 1727204083.62404: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204083.3490949-15060-59900587035213/ > /dev/null 2>&1 && sleep 0' 13118 1727204083.64530: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204083.64544: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.64553: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.64570: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.64619: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.64626: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204083.64640: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.64654: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204083.64679: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204083.64687: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204083.64702: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.64711: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.64723: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.64813: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.64820: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204083.64830: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.64905: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204083.65029: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204083.65046: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204083.65141: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204083.67043: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204083.67047: stdout chunk (state=3): >>><<< 13118 1727204083.67050: stderr chunk (state=3): >>><<< 13118 1727204083.67176: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204083.67180: handler run complete 13118 1727204083.67183: attempt loop complete, returning result 13118 1727204083.67185: _execute() done 13118 1727204083.67187: dumping result to json 13118 1727204083.67189: done dumping result, returning 13118 1727204083.67191: done running TaskExecutor() for managed-node2/TASK: Stat profile file [0affcd87-79f5-56a3-0a64-0000000003b1] 13118 1727204083.67193: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b1 13118 1727204083.67328: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b1 ok: [managed-node2] => { "changed": false, "stat": { "exists": false } } 13118 1727204083.67392: no more pending results, returning what we have 13118 1727204083.67397: results queue empty 13118 1727204083.67398: checking for any_errors_fatal 13118 1727204083.67405: done checking for any_errors_fatal 13118 1727204083.67406: checking for max_fail_percentage 13118 1727204083.67408: done checking for max_fail_percentage 13118 1727204083.67410: checking to see if all hosts have failed and the running result is not ok 13118 1727204083.67410: done checking to see if all hosts have failed 13118 1727204083.67411: getting the remaining hosts for this loop 13118 1727204083.67413: done getting the remaining hosts for this loop 13118 1727204083.67416: getting the next task for host managed-node2 13118 1727204083.67425: done getting next task for host managed-node2 13118 1727204083.67427: ^ task is: TASK: Set NM profile exist flag based on the profile files 13118 1727204083.67434: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204083.67439: getting variables 13118 1727204083.67440: in VariableManager get_vars() 13118 1727204083.67511: Calling all_inventory to load vars for managed-node2 13118 1727204083.67514: Calling groups_inventory to load vars for managed-node2 13118 1727204083.67517: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204083.67534: Calling all_plugins_play to load vars for managed-node2 13118 1727204083.67538: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204083.67542: Calling groups_plugins_play to load vars for managed-node2 13118 1727204083.68454: WORKER PROCESS EXITING 13118 1727204083.70336: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204083.71273: done with get_vars() 13118 1727204083.71294: done getting variables 13118 1727204083.71344: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Set NM profile exist flag based on the profile files] ******************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:17 Tuesday 24 September 2024 14:54:43 -0400 (0:00:00.412) 0:00:21.013 ***** 13118 1727204083.71370: entering _queue_task() for managed-node2/set_fact 13118 1727204083.71611: worker is 1 (out of 1 available) 13118 1727204083.71625: exiting _queue_task() for managed-node2/set_fact 13118 1727204083.71636: done queuing things up, now waiting for results queue to drain 13118 1727204083.71638: waiting for pending results... 13118 1727204083.71919: running TaskExecutor() for managed-node2/TASK: Set NM profile exist flag based on the profile files 13118 1727204083.72204: in run() - task 0affcd87-79f5-56a3-0a64-0000000003b2 13118 1727204083.72226: variable 'ansible_search_path' from source: unknown 13118 1727204083.72235: variable 'ansible_search_path' from source: unknown 13118 1727204083.72279: calling self._execute() 13118 1727204083.72463: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.72478: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.72493: variable 'omit' from source: magic vars 13118 1727204083.72858: variable 'ansible_distribution_major_version' from source: facts 13118 1727204083.73420: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204083.73553: variable 'profile_stat' from source: set_fact 13118 1727204083.73575: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204083.73584: when evaluation is False, skipping this task 13118 1727204083.73593: _execute() done 13118 1727204083.73600: dumping result to json 13118 1727204083.73609: done dumping result, returning 13118 1727204083.73620: done running TaskExecutor() for managed-node2/TASK: Set NM profile exist flag based on the profile files [0affcd87-79f5-56a3-0a64-0000000003b2] 13118 1727204083.73631: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b2 skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204083.73784: no more pending results, returning what we have 13118 1727204083.73789: results queue empty 13118 1727204083.73790: checking for any_errors_fatal 13118 1727204083.73798: done checking for any_errors_fatal 13118 1727204083.73799: checking for max_fail_percentage 13118 1727204083.73800: done checking for max_fail_percentage 13118 1727204083.73801: checking to see if all hosts have failed and the running result is not ok 13118 1727204083.73802: done checking to see if all hosts have failed 13118 1727204083.73803: getting the remaining hosts for this loop 13118 1727204083.73804: done getting the remaining hosts for this loop 13118 1727204083.73808: getting the next task for host managed-node2 13118 1727204083.73815: done getting next task for host managed-node2 13118 1727204083.73817: ^ task is: TASK: Get NM profile info 13118 1727204083.73821: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204083.73825: getting variables 13118 1727204083.73827: in VariableManager get_vars() 13118 1727204083.73874: Calling all_inventory to load vars for managed-node2 13118 1727204083.73877: Calling groups_inventory to load vars for managed-node2 13118 1727204083.73880: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204083.73888: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b2 13118 1727204083.73893: WORKER PROCESS EXITING 13118 1727204083.73907: Calling all_plugins_play to load vars for managed-node2 13118 1727204083.73911: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204083.73915: Calling groups_plugins_play to load vars for managed-node2 13118 1727204083.76266: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204083.80334: done with get_vars() 13118 1727204083.80382: done getting variables 13118 1727204083.80505: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Get NM profile info] ***************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:25 Tuesday 24 September 2024 14:54:43 -0400 (0:00:00.091) 0:00:21.105 ***** 13118 1727204083.80570: entering _queue_task() for managed-node2/shell 13118 1727204083.81082: worker is 1 (out of 1 available) 13118 1727204083.81098: exiting _queue_task() for managed-node2/shell 13118 1727204083.81109: done queuing things up, now waiting for results queue to drain 13118 1727204083.81110: waiting for pending results... 13118 1727204083.81411: running TaskExecutor() for managed-node2/TASK: Get NM profile info 13118 1727204083.81540: in run() - task 0affcd87-79f5-56a3-0a64-0000000003b3 13118 1727204083.81557: variable 'ansible_search_path' from source: unknown 13118 1727204083.81560: variable 'ansible_search_path' from source: unknown 13118 1727204083.81599: calling self._execute() 13118 1727204083.81708: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.81713: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.81723: variable 'omit' from source: magic vars 13118 1727204083.82153: variable 'ansible_distribution_major_version' from source: facts 13118 1727204083.82171: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204083.82182: variable 'omit' from source: magic vars 13118 1727204083.82235: variable 'omit' from source: magic vars 13118 1727204083.82353: variable 'profile' from source: include params 13118 1727204083.82389: variable 'item' from source: include params 13118 1727204083.82470: variable 'item' from source: include params 13118 1727204083.82503: variable 'omit' from source: magic vars 13118 1727204083.82851: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204083.82855: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204083.83118: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204083.83122: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204083.83125: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204083.83128: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204083.83130: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.83132: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.83135: Set connection var ansible_timeout to 10 13118 1727204083.83137: Set connection var ansible_pipelining to False 13118 1727204083.83140: Set connection var ansible_connection to ssh 13118 1727204083.83142: Set connection var ansible_shell_executable to /bin/sh 13118 1727204083.83144: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204083.83146: Set connection var ansible_shell_type to sh 13118 1727204083.83148: variable 'ansible_shell_executable' from source: unknown 13118 1727204083.83151: variable 'ansible_connection' from source: unknown 13118 1727204083.83154: variable 'ansible_module_compression' from source: unknown 13118 1727204083.83156: variable 'ansible_shell_type' from source: unknown 13118 1727204083.83158: variable 'ansible_shell_executable' from source: unknown 13118 1727204083.83161: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204083.83163: variable 'ansible_pipelining' from source: unknown 13118 1727204083.83167: variable 'ansible_timeout' from source: unknown 13118 1727204083.83169: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204083.83172: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204083.83174: variable 'omit' from source: magic vars 13118 1727204083.83176: starting attempt loop 13118 1727204083.83178: running the handler 13118 1727204083.83180: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204083.83183: _low_level_execute_command(): starting 13118 1727204083.83468: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204083.84291: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204083.84295: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.84302: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.84305: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.84307: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.84309: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204083.84311: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.84313: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204083.84315: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204083.84317: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204083.84319: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.84320: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.84322: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.84324: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.84326: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204083.84328: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.84330: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204083.84331: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204083.84333: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204083.84469: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204083.86071: stdout chunk (state=3): >>>/root <<< 13118 1727204083.86239: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204083.86242: stdout chunk (state=3): >>><<< 13118 1727204083.86252: stderr chunk (state=3): >>><<< 13118 1727204083.86312: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204083.86336: _low_level_execute_command(): starting 13118 1727204083.86340: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204083.8631237-15089-258516884008520 `" && echo ansible-tmp-1727204083.8631237-15089-258516884008520="` echo /root/.ansible/tmp/ansible-tmp-1727204083.8631237-15089-258516884008520 `" ) && sleep 0' 13118 1727204083.87071: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204083.87088: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.87103: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.87122: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.87172: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.87186: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204083.87201: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.87220: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204083.87235: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204083.87248: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204083.87259: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.87277: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.87294: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.87307: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.87319: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204083.87337: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.87415: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204083.87441: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204083.87458: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204083.87541: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204083.89438: stdout chunk (state=3): >>>ansible-tmp-1727204083.8631237-15089-258516884008520=/root/.ansible/tmp/ansible-tmp-1727204083.8631237-15089-258516884008520 <<< 13118 1727204083.89671: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204083.89675: stdout chunk (state=3): >>><<< 13118 1727204083.89677: stderr chunk (state=3): >>><<< 13118 1727204083.89680: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204083.8631237-15089-258516884008520=/root/.ansible/tmp/ansible-tmp-1727204083.8631237-15089-258516884008520 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204083.89682: variable 'ansible_module_compression' from source: unknown 13118 1727204083.89933: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13118 1727204083.89937: variable 'ansible_facts' from source: unknown 13118 1727204083.89941: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204083.8631237-15089-258516884008520/AnsiballZ_command.py 13118 1727204083.90006: Sending initial data 13118 1727204083.90010: Sent initial data (156 bytes) 13118 1727204083.90902: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204083.90916: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.90930: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.90947: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.90990: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.91001: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204083.91013: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.91028: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204083.91039: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204083.91047: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204083.91058: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.91075: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.91091: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.91103: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.91114: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204083.91127: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.91195: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204083.91213: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204083.91227: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204083.91294: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204083.93030: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204083.93113: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204083.93117: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmp3mt6j91_ /root/.ansible/tmp/ansible-tmp-1727204083.8631237-15089-258516884008520/AnsiballZ_command.py <<< 13118 1727204083.93148: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204083.94205: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204083.94502: stderr chunk (state=3): >>><<< 13118 1727204083.94506: stdout chunk (state=3): >>><<< 13118 1727204083.94509: done transferring module to remote 13118 1727204083.94511: _low_level_execute_command(): starting 13118 1727204083.94514: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204083.8631237-15089-258516884008520/ /root/.ansible/tmp/ansible-tmp-1727204083.8631237-15089-258516884008520/AnsiballZ_command.py && sleep 0' 13118 1727204083.95178: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204083.95193: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.95209: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.95227: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.95276: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.95295: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204083.95310: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.95329: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204083.95342: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204083.95353: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204083.95368: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.95386: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.95408: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.95421: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.95431: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204083.95445: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.95532: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204083.95554: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204083.95573: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204083.95647: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204083.97370: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204083.97475: stderr chunk (state=3): >>><<< 13118 1727204083.97492: stdout chunk (state=3): >>><<< 13118 1727204083.97612: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204083.97616: _low_level_execute_command(): starting 13118 1727204083.97619: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204083.8631237-15089-258516884008520/AnsiballZ_command.py && sleep 0' 13118 1727204083.98225: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204083.98243: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.98267: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.98286: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.98328: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.98344: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204083.98359: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.98385: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204083.98397: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204083.98407: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204083.98419: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204083.98433: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204083.98448: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204083.98459: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204083.98471: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204083.98491: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204083.98569: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204083.98599: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204083.98616: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204083.98694: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204084.14922: stdout chunk (state=3): >>> {"changed": true, "stdout": "bond0 /etc/NetworkManager/system-connections/bond0.nmconnection \nbond0.0 /etc/NetworkManager/system-connections/bond0.0.nmconnection \nbond0.1 /etc/NetworkManager/system-connections/bond0.1.nmconnection ", "stderr": "", "rc": 0, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0 | grep /etc", "start": "2024-09-24 14:54:44.118851", "end": "2024-09-24 14:54:44.148379", "delta": "0:00:00.029528", "msg": "", "invocation": {"module_args": {"_raw_params": "nmcli -f NAME,FILENAME connection show |grep bond0 | grep /etc", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13118 1727204084.16222: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204084.16286: stderr chunk (state=3): >>><<< 13118 1727204084.16290: stdout chunk (state=3): >>><<< 13118 1727204084.16305: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "bond0 /etc/NetworkManager/system-connections/bond0.nmconnection \nbond0.0 /etc/NetworkManager/system-connections/bond0.0.nmconnection \nbond0.1 /etc/NetworkManager/system-connections/bond0.1.nmconnection ", "stderr": "", "rc": 0, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0 | grep /etc", "start": "2024-09-24 14:54:44.118851", "end": "2024-09-24 14:54:44.148379", "delta": "0:00:00.029528", "msg": "", "invocation": {"module_args": {"_raw_params": "nmcli -f NAME,FILENAME connection show |grep bond0 | grep /etc", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204084.16341: done with _execute_module (ansible.legacy.command, {'_raw_params': 'nmcli -f NAME,FILENAME connection show |grep bond0 | grep /etc', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204083.8631237-15089-258516884008520/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204084.16350: _low_level_execute_command(): starting 13118 1727204084.16355: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204083.8631237-15089-258516884008520/ > /dev/null 2>&1 && sleep 0' 13118 1727204084.16851: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204084.16855: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204084.16902: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204084.16906: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204084.16908: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204084.16959: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204084.16962: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204084.16966: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204084.17013: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204084.18820: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204084.18882: stderr chunk (state=3): >>><<< 13118 1727204084.18885: stdout chunk (state=3): >>><<< 13118 1727204084.18900: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204084.18907: handler run complete 13118 1727204084.18925: Evaluated conditional (False): False 13118 1727204084.18938: attempt loop complete, returning result 13118 1727204084.18941: _execute() done 13118 1727204084.18943: dumping result to json 13118 1727204084.18945: done dumping result, returning 13118 1727204084.18955: done running TaskExecutor() for managed-node2/TASK: Get NM profile info [0affcd87-79f5-56a3-0a64-0000000003b3] 13118 1727204084.18961: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b3 13118 1727204084.19059: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b3 13118 1727204084.19062: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0 | grep /etc", "delta": "0:00:00.029528", "end": "2024-09-24 14:54:44.148379", "rc": 0, "start": "2024-09-24 14:54:44.118851" } STDOUT: bond0 /etc/NetworkManager/system-connections/bond0.nmconnection bond0.0 /etc/NetworkManager/system-connections/bond0.0.nmconnection bond0.1 /etc/NetworkManager/system-connections/bond0.1.nmconnection 13118 1727204084.19133: no more pending results, returning what we have 13118 1727204084.19137: results queue empty 13118 1727204084.19138: checking for any_errors_fatal 13118 1727204084.19143: done checking for any_errors_fatal 13118 1727204084.19143: checking for max_fail_percentage 13118 1727204084.19145: done checking for max_fail_percentage 13118 1727204084.19146: checking to see if all hosts have failed and the running result is not ok 13118 1727204084.19147: done checking to see if all hosts have failed 13118 1727204084.19147: getting the remaining hosts for this loop 13118 1727204084.19149: done getting the remaining hosts for this loop 13118 1727204084.19152: getting the next task for host managed-node2 13118 1727204084.19160: done getting next task for host managed-node2 13118 1727204084.19162: ^ task is: TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output 13118 1727204084.19168: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204084.19171: getting variables 13118 1727204084.19173: in VariableManager get_vars() 13118 1727204084.19214: Calling all_inventory to load vars for managed-node2 13118 1727204084.19217: Calling groups_inventory to load vars for managed-node2 13118 1727204084.19219: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204084.19232: Calling all_plugins_play to load vars for managed-node2 13118 1727204084.19235: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204084.19237: Calling groups_plugins_play to load vars for managed-node2 13118 1727204084.20066: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204084.20997: done with get_vars() 13118 1727204084.21014: done getting variables 13118 1727204084.21063: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Set NM profile exist flag and ansible_managed flag true based on the nmcli output] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:35 Tuesday 24 September 2024 14:54:44 -0400 (0:00:00.405) 0:00:21.510 ***** 13118 1727204084.21092: entering _queue_task() for managed-node2/set_fact 13118 1727204084.21324: worker is 1 (out of 1 available) 13118 1727204084.21341: exiting _queue_task() for managed-node2/set_fact 13118 1727204084.21353: done queuing things up, now waiting for results queue to drain 13118 1727204084.21354: waiting for pending results... 13118 1727204084.21535: running TaskExecutor() for managed-node2/TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output 13118 1727204084.21609: in run() - task 0affcd87-79f5-56a3-0a64-0000000003b4 13118 1727204084.21620: variable 'ansible_search_path' from source: unknown 13118 1727204084.21624: variable 'ansible_search_path' from source: unknown 13118 1727204084.21655: calling self._execute() 13118 1727204084.21727: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.21731: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.21742: variable 'omit' from source: magic vars 13118 1727204084.22026: variable 'ansible_distribution_major_version' from source: facts 13118 1727204084.22039: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204084.22128: variable 'nm_profile_exists' from source: set_fact 13118 1727204084.22141: Evaluated conditional (nm_profile_exists.rc == 0): True 13118 1727204084.22146: variable 'omit' from source: magic vars 13118 1727204084.22178: variable 'omit' from source: magic vars 13118 1727204084.22200: variable 'omit' from source: magic vars 13118 1727204084.22237: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204084.22263: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204084.22281: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204084.22294: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204084.22304: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204084.22327: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204084.22330: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.22336: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.22405: Set connection var ansible_timeout to 10 13118 1727204084.22415: Set connection var ansible_pipelining to False 13118 1727204084.22418: Set connection var ansible_connection to ssh 13118 1727204084.22423: Set connection var ansible_shell_executable to /bin/sh 13118 1727204084.22428: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204084.22431: Set connection var ansible_shell_type to sh 13118 1727204084.22454: variable 'ansible_shell_executable' from source: unknown 13118 1727204084.22459: variable 'ansible_connection' from source: unknown 13118 1727204084.22461: variable 'ansible_module_compression' from source: unknown 13118 1727204084.22464: variable 'ansible_shell_type' from source: unknown 13118 1727204084.22466: variable 'ansible_shell_executable' from source: unknown 13118 1727204084.22469: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.22472: variable 'ansible_pipelining' from source: unknown 13118 1727204084.22474: variable 'ansible_timeout' from source: unknown 13118 1727204084.22476: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.22577: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204084.22586: variable 'omit' from source: magic vars 13118 1727204084.22591: starting attempt loop 13118 1727204084.22594: running the handler 13118 1727204084.22606: handler run complete 13118 1727204084.22614: attempt loop complete, returning result 13118 1727204084.22617: _execute() done 13118 1727204084.22620: dumping result to json 13118 1727204084.22622: done dumping result, returning 13118 1727204084.22628: done running TaskExecutor() for managed-node2/TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output [0affcd87-79f5-56a3-0a64-0000000003b4] 13118 1727204084.22637: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b4 13118 1727204084.22712: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b4 13118 1727204084.22715: WORKER PROCESS EXITING ok: [managed-node2] => { "ansible_facts": { "lsr_net_profile_ansible_managed": true, "lsr_net_profile_exists": true, "lsr_net_profile_fingerprint": true }, "changed": false } 13118 1727204084.22768: no more pending results, returning what we have 13118 1727204084.22772: results queue empty 13118 1727204084.22773: checking for any_errors_fatal 13118 1727204084.22779: done checking for any_errors_fatal 13118 1727204084.22780: checking for max_fail_percentage 13118 1727204084.22782: done checking for max_fail_percentage 13118 1727204084.22783: checking to see if all hosts have failed and the running result is not ok 13118 1727204084.22783: done checking to see if all hosts have failed 13118 1727204084.22784: getting the remaining hosts for this loop 13118 1727204084.22785: done getting the remaining hosts for this loop 13118 1727204084.22789: getting the next task for host managed-node2 13118 1727204084.22799: done getting next task for host managed-node2 13118 1727204084.22801: ^ task is: TASK: Get the ansible_managed comment in ifcfg-{{ profile }} 13118 1727204084.22805: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204084.22810: getting variables 13118 1727204084.22811: in VariableManager get_vars() 13118 1727204084.22860: Calling all_inventory to load vars for managed-node2 13118 1727204084.22862: Calling groups_inventory to load vars for managed-node2 13118 1727204084.22866: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204084.22876: Calling all_plugins_play to load vars for managed-node2 13118 1727204084.22878: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204084.22881: Calling groups_plugins_play to load vars for managed-node2 13118 1727204084.23818: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204084.24722: done with get_vars() 13118 1727204084.24738: done getting variables 13118 1727204084.24784: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204084.24875: variable 'profile' from source: include params 13118 1727204084.24878: variable 'item' from source: include params 13118 1727204084.24921: variable 'item' from source: include params TASK [Get the ansible_managed comment in ifcfg-bond0] ************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:49 Tuesday 24 September 2024 14:54:44 -0400 (0:00:00.038) 0:00:21.549 ***** 13118 1727204084.24949: entering _queue_task() for managed-node2/command 13118 1727204084.25172: worker is 1 (out of 1 available) 13118 1727204084.25185: exiting _queue_task() for managed-node2/command 13118 1727204084.25196: done queuing things up, now waiting for results queue to drain 13118 1727204084.25197: waiting for pending results... 13118 1727204084.25376: running TaskExecutor() for managed-node2/TASK: Get the ansible_managed comment in ifcfg-bond0 13118 1727204084.25462: in run() - task 0affcd87-79f5-56a3-0a64-0000000003b6 13118 1727204084.25473: variable 'ansible_search_path' from source: unknown 13118 1727204084.25477: variable 'ansible_search_path' from source: unknown 13118 1727204084.25503: calling self._execute() 13118 1727204084.25575: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.25578: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.25587: variable 'omit' from source: magic vars 13118 1727204084.25850: variable 'ansible_distribution_major_version' from source: facts 13118 1727204084.25859: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204084.25943: variable 'profile_stat' from source: set_fact 13118 1727204084.25955: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204084.25958: when evaluation is False, skipping this task 13118 1727204084.25962: _execute() done 13118 1727204084.25965: dumping result to json 13118 1727204084.25968: done dumping result, returning 13118 1727204084.25977: done running TaskExecutor() for managed-node2/TASK: Get the ansible_managed comment in ifcfg-bond0 [0affcd87-79f5-56a3-0a64-0000000003b6] 13118 1727204084.25980: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b6 13118 1727204084.26062: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b6 13118 1727204084.26067: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204084.26128: no more pending results, returning what we have 13118 1727204084.26132: results queue empty 13118 1727204084.26133: checking for any_errors_fatal 13118 1727204084.26139: done checking for any_errors_fatal 13118 1727204084.26140: checking for max_fail_percentage 13118 1727204084.26141: done checking for max_fail_percentage 13118 1727204084.26142: checking to see if all hosts have failed and the running result is not ok 13118 1727204084.26143: done checking to see if all hosts have failed 13118 1727204084.26144: getting the remaining hosts for this loop 13118 1727204084.26145: done getting the remaining hosts for this loop 13118 1727204084.26149: getting the next task for host managed-node2 13118 1727204084.26155: done getting next task for host managed-node2 13118 1727204084.26157: ^ task is: TASK: Verify the ansible_managed comment in ifcfg-{{ profile }} 13118 1727204084.26161: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204084.26166: getting variables 13118 1727204084.26168: in VariableManager get_vars() 13118 1727204084.26211: Calling all_inventory to load vars for managed-node2 13118 1727204084.26213: Calling groups_inventory to load vars for managed-node2 13118 1727204084.26215: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204084.26225: Calling all_plugins_play to load vars for managed-node2 13118 1727204084.26227: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204084.26229: Calling groups_plugins_play to load vars for managed-node2 13118 1727204084.27016: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204084.27941: done with get_vars() 13118 1727204084.27958: done getting variables 13118 1727204084.28003: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204084.28089: variable 'profile' from source: include params 13118 1727204084.28092: variable 'item' from source: include params 13118 1727204084.28131: variable 'item' from source: include params TASK [Verify the ansible_managed comment in ifcfg-bond0] *********************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:56 Tuesday 24 September 2024 14:54:44 -0400 (0:00:00.032) 0:00:21.581 ***** 13118 1727204084.28156: entering _queue_task() for managed-node2/set_fact 13118 1727204084.28386: worker is 1 (out of 1 available) 13118 1727204084.28399: exiting _queue_task() for managed-node2/set_fact 13118 1727204084.28412: done queuing things up, now waiting for results queue to drain 13118 1727204084.28413: waiting for pending results... 13118 1727204084.28599: running TaskExecutor() for managed-node2/TASK: Verify the ansible_managed comment in ifcfg-bond0 13118 1727204084.28679: in run() - task 0affcd87-79f5-56a3-0a64-0000000003b7 13118 1727204084.28693: variable 'ansible_search_path' from source: unknown 13118 1727204084.28696: variable 'ansible_search_path' from source: unknown 13118 1727204084.28720: calling self._execute() 13118 1727204084.28798: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.28802: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.28809: variable 'omit' from source: magic vars 13118 1727204084.29076: variable 'ansible_distribution_major_version' from source: facts 13118 1727204084.29086: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204084.29169: variable 'profile_stat' from source: set_fact 13118 1727204084.29182: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204084.29185: when evaluation is False, skipping this task 13118 1727204084.29188: _execute() done 13118 1727204084.29190: dumping result to json 13118 1727204084.29193: done dumping result, returning 13118 1727204084.29199: done running TaskExecutor() for managed-node2/TASK: Verify the ansible_managed comment in ifcfg-bond0 [0affcd87-79f5-56a3-0a64-0000000003b7] 13118 1727204084.29205: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b7 13118 1727204084.29296: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b7 13118 1727204084.29299: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204084.29349: no more pending results, returning what we have 13118 1727204084.29353: results queue empty 13118 1727204084.29354: checking for any_errors_fatal 13118 1727204084.29360: done checking for any_errors_fatal 13118 1727204084.29360: checking for max_fail_percentage 13118 1727204084.29362: done checking for max_fail_percentage 13118 1727204084.29363: checking to see if all hosts have failed and the running result is not ok 13118 1727204084.29365: done checking to see if all hosts have failed 13118 1727204084.29366: getting the remaining hosts for this loop 13118 1727204084.29367: done getting the remaining hosts for this loop 13118 1727204084.29371: getting the next task for host managed-node2 13118 1727204084.29379: done getting next task for host managed-node2 13118 1727204084.29382: ^ task is: TASK: Get the fingerprint comment in ifcfg-{{ profile }} 13118 1727204084.29386: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204084.29391: getting variables 13118 1727204084.29392: in VariableManager get_vars() 13118 1727204084.29429: Calling all_inventory to load vars for managed-node2 13118 1727204084.29431: Calling groups_inventory to load vars for managed-node2 13118 1727204084.29433: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204084.29450: Calling all_plugins_play to load vars for managed-node2 13118 1727204084.29453: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204084.29456: Calling groups_plugins_play to load vars for managed-node2 13118 1727204084.30388: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204084.31302: done with get_vars() 13118 1727204084.31321: done getting variables 13118 1727204084.31369: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204084.31455: variable 'profile' from source: include params 13118 1727204084.31458: variable 'item' from source: include params 13118 1727204084.31501: variable 'item' from source: include params TASK [Get the fingerprint comment in ifcfg-bond0] ****************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:62 Tuesday 24 September 2024 14:54:44 -0400 (0:00:00.033) 0:00:21.615 ***** 13118 1727204084.31527: entering _queue_task() for managed-node2/command 13118 1727204084.31766: worker is 1 (out of 1 available) 13118 1727204084.31781: exiting _queue_task() for managed-node2/command 13118 1727204084.31793: done queuing things up, now waiting for results queue to drain 13118 1727204084.31794: waiting for pending results... 13118 1727204084.31987: running TaskExecutor() for managed-node2/TASK: Get the fingerprint comment in ifcfg-bond0 13118 1727204084.32071: in run() - task 0affcd87-79f5-56a3-0a64-0000000003b8 13118 1727204084.32083: variable 'ansible_search_path' from source: unknown 13118 1727204084.32086: variable 'ansible_search_path' from source: unknown 13118 1727204084.32114: calling self._execute() 13118 1727204084.32191: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.32195: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.32203: variable 'omit' from source: magic vars 13118 1727204084.32478: variable 'ansible_distribution_major_version' from source: facts 13118 1727204084.32490: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204084.32573: variable 'profile_stat' from source: set_fact 13118 1727204084.32584: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204084.32589: when evaluation is False, skipping this task 13118 1727204084.32592: _execute() done 13118 1727204084.32595: dumping result to json 13118 1727204084.32597: done dumping result, returning 13118 1727204084.32601: done running TaskExecutor() for managed-node2/TASK: Get the fingerprint comment in ifcfg-bond0 [0affcd87-79f5-56a3-0a64-0000000003b8] 13118 1727204084.32609: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b8 13118 1727204084.32690: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b8 13118 1727204084.32693: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204084.32747: no more pending results, returning what we have 13118 1727204084.32752: results queue empty 13118 1727204084.32753: checking for any_errors_fatal 13118 1727204084.32760: done checking for any_errors_fatal 13118 1727204084.32761: checking for max_fail_percentage 13118 1727204084.32762: done checking for max_fail_percentage 13118 1727204084.32763: checking to see if all hosts have failed and the running result is not ok 13118 1727204084.32765: done checking to see if all hosts have failed 13118 1727204084.32766: getting the remaining hosts for this loop 13118 1727204084.32767: done getting the remaining hosts for this loop 13118 1727204084.32771: getting the next task for host managed-node2 13118 1727204084.32778: done getting next task for host managed-node2 13118 1727204084.32781: ^ task is: TASK: Verify the fingerprint comment in ifcfg-{{ profile }} 13118 1727204084.32784: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204084.32790: getting variables 13118 1727204084.32793: in VariableManager get_vars() 13118 1727204084.32840: Calling all_inventory to load vars for managed-node2 13118 1727204084.32842: Calling groups_inventory to load vars for managed-node2 13118 1727204084.32844: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204084.32855: Calling all_plugins_play to load vars for managed-node2 13118 1727204084.32857: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204084.32860: Calling groups_plugins_play to load vars for managed-node2 13118 1727204084.33675: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204084.34594: done with get_vars() 13118 1727204084.34612: done getting variables 13118 1727204084.34656: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204084.34742: variable 'profile' from source: include params 13118 1727204084.34745: variable 'item' from source: include params 13118 1727204084.34788: variable 'item' from source: include params TASK [Verify the fingerprint comment in ifcfg-bond0] *************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:69 Tuesday 24 September 2024 14:54:44 -0400 (0:00:00.032) 0:00:21.648 ***** 13118 1727204084.34812: entering _queue_task() for managed-node2/set_fact 13118 1727204084.35039: worker is 1 (out of 1 available) 13118 1727204084.35053: exiting _queue_task() for managed-node2/set_fact 13118 1727204084.35066: done queuing things up, now waiting for results queue to drain 13118 1727204084.35067: waiting for pending results... 13118 1727204084.35514: running TaskExecutor() for managed-node2/TASK: Verify the fingerprint comment in ifcfg-bond0 13118 1727204084.35595: in run() - task 0affcd87-79f5-56a3-0a64-0000000003b9 13118 1727204084.35605: variable 'ansible_search_path' from source: unknown 13118 1727204084.35608: variable 'ansible_search_path' from source: unknown 13118 1727204084.35639: calling self._execute() 13118 1727204084.35713: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.35718: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.35726: variable 'omit' from source: magic vars 13118 1727204084.35991: variable 'ansible_distribution_major_version' from source: facts 13118 1727204084.36002: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204084.36087: variable 'profile_stat' from source: set_fact 13118 1727204084.36097: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204084.36101: when evaluation is False, skipping this task 13118 1727204084.36104: _execute() done 13118 1727204084.36106: dumping result to json 13118 1727204084.36109: done dumping result, returning 13118 1727204084.36119: done running TaskExecutor() for managed-node2/TASK: Verify the fingerprint comment in ifcfg-bond0 [0affcd87-79f5-56a3-0a64-0000000003b9] 13118 1727204084.36122: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b9 13118 1727204084.36207: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003b9 13118 1727204084.36209: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204084.36292: no more pending results, returning what we have 13118 1727204084.36296: results queue empty 13118 1727204084.36297: checking for any_errors_fatal 13118 1727204084.36302: done checking for any_errors_fatal 13118 1727204084.36302: checking for max_fail_percentage 13118 1727204084.36304: done checking for max_fail_percentage 13118 1727204084.36305: checking to see if all hosts have failed and the running result is not ok 13118 1727204084.36306: done checking to see if all hosts have failed 13118 1727204084.36306: getting the remaining hosts for this loop 13118 1727204084.36307: done getting the remaining hosts for this loop 13118 1727204084.36311: getting the next task for host managed-node2 13118 1727204084.36318: done getting next task for host managed-node2 13118 1727204084.36320: ^ task is: TASK: Assert that the profile is present - '{{ profile }}' 13118 1727204084.36323: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204084.36326: getting variables 13118 1727204084.36327: in VariableManager get_vars() 13118 1727204084.36369: Calling all_inventory to load vars for managed-node2 13118 1727204084.36372: Calling groups_inventory to load vars for managed-node2 13118 1727204084.36374: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204084.36384: Calling all_plugins_play to load vars for managed-node2 13118 1727204084.36386: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204084.36389: Calling groups_plugins_play to load vars for managed-node2 13118 1727204084.37595: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204084.39274: done with get_vars() 13118 1727204084.39301: done getting variables 13118 1727204084.39367: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204084.39488: variable 'profile' from source: include params 13118 1727204084.39492: variable 'item' from source: include params 13118 1727204084.39554: variable 'item' from source: include params TASK [Assert that the profile is present - 'bond0'] **************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:5 Tuesday 24 September 2024 14:54:44 -0400 (0:00:00.047) 0:00:21.695 ***** 13118 1727204084.39586: entering _queue_task() for managed-node2/assert 13118 1727204084.39928: worker is 1 (out of 1 available) 13118 1727204084.39942: exiting _queue_task() for managed-node2/assert 13118 1727204084.39954: done queuing things up, now waiting for results queue to drain 13118 1727204084.39955: waiting for pending results... 13118 1727204084.40250: running TaskExecutor() for managed-node2/TASK: Assert that the profile is present - 'bond0' 13118 1727204084.40366: in run() - task 0affcd87-79f5-56a3-0a64-000000000260 13118 1727204084.40386: variable 'ansible_search_path' from source: unknown 13118 1727204084.40395: variable 'ansible_search_path' from source: unknown 13118 1727204084.40439: calling self._execute() 13118 1727204084.40573: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.40584: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.40613: variable 'omit' from source: magic vars 13118 1727204084.41037: variable 'ansible_distribution_major_version' from source: facts 13118 1727204084.41055: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204084.41065: variable 'omit' from source: magic vars 13118 1727204084.41099: variable 'omit' from source: magic vars 13118 1727204084.41172: variable 'profile' from source: include params 13118 1727204084.41175: variable 'item' from source: include params 13118 1727204084.41223: variable 'item' from source: include params 13118 1727204084.41238: variable 'omit' from source: magic vars 13118 1727204084.41275: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204084.41306: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204084.41323: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204084.41337: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204084.41346: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204084.41371: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204084.41375: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.41377: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.41450: Set connection var ansible_timeout to 10 13118 1727204084.41459: Set connection var ansible_pipelining to False 13118 1727204084.41462: Set connection var ansible_connection to ssh 13118 1727204084.41469: Set connection var ansible_shell_executable to /bin/sh 13118 1727204084.41474: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204084.41477: Set connection var ansible_shell_type to sh 13118 1727204084.41497: variable 'ansible_shell_executable' from source: unknown 13118 1727204084.41500: variable 'ansible_connection' from source: unknown 13118 1727204084.41503: variable 'ansible_module_compression' from source: unknown 13118 1727204084.41505: variable 'ansible_shell_type' from source: unknown 13118 1727204084.41507: variable 'ansible_shell_executable' from source: unknown 13118 1727204084.41511: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.41513: variable 'ansible_pipelining' from source: unknown 13118 1727204084.41515: variable 'ansible_timeout' from source: unknown 13118 1727204084.41517: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.41618: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204084.41628: variable 'omit' from source: magic vars 13118 1727204084.41634: starting attempt loop 13118 1727204084.41638: running the handler 13118 1727204084.41716: variable 'lsr_net_profile_exists' from source: set_fact 13118 1727204084.41721: Evaluated conditional (lsr_net_profile_exists): True 13118 1727204084.41727: handler run complete 13118 1727204084.41739: attempt loop complete, returning result 13118 1727204084.41742: _execute() done 13118 1727204084.41745: dumping result to json 13118 1727204084.41749: done dumping result, returning 13118 1727204084.41757: done running TaskExecutor() for managed-node2/TASK: Assert that the profile is present - 'bond0' [0affcd87-79f5-56a3-0a64-000000000260] 13118 1727204084.41759: sending task result for task 0affcd87-79f5-56a3-0a64-000000000260 13118 1727204084.41847: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000260 13118 1727204084.41850: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false } MSG: All assertions passed 13118 1727204084.41896: no more pending results, returning what we have 13118 1727204084.41900: results queue empty 13118 1727204084.41901: checking for any_errors_fatal 13118 1727204084.41906: done checking for any_errors_fatal 13118 1727204084.41907: checking for max_fail_percentage 13118 1727204084.41909: done checking for max_fail_percentage 13118 1727204084.41910: checking to see if all hosts have failed and the running result is not ok 13118 1727204084.41910: done checking to see if all hosts have failed 13118 1727204084.41911: getting the remaining hosts for this loop 13118 1727204084.41912: done getting the remaining hosts for this loop 13118 1727204084.41916: getting the next task for host managed-node2 13118 1727204084.41922: done getting next task for host managed-node2 13118 1727204084.41925: ^ task is: TASK: Assert that the ansible managed comment is present in '{{ profile }}' 13118 1727204084.41928: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204084.41934: getting variables 13118 1727204084.41936: in VariableManager get_vars() 13118 1727204084.41985: Calling all_inventory to load vars for managed-node2 13118 1727204084.41988: Calling groups_inventory to load vars for managed-node2 13118 1727204084.41990: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204084.42000: Calling all_plugins_play to load vars for managed-node2 13118 1727204084.42002: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204084.42004: Calling groups_plugins_play to load vars for managed-node2 13118 1727204084.43015: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204084.44972: done with get_vars() 13118 1727204084.44995: done getting variables 13118 1727204084.45065: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204084.45194: variable 'profile' from source: include params 13118 1727204084.45198: variable 'item' from source: include params 13118 1727204084.45268: variable 'item' from source: include params TASK [Assert that the ansible managed comment is present in 'bond0'] *********** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:10 Tuesday 24 September 2024 14:54:44 -0400 (0:00:00.057) 0:00:21.752 ***** 13118 1727204084.45305: entering _queue_task() for managed-node2/assert 13118 1727204084.45688: worker is 1 (out of 1 available) 13118 1727204084.45700: exiting _queue_task() for managed-node2/assert 13118 1727204084.45712: done queuing things up, now waiting for results queue to drain 13118 1727204084.45713: waiting for pending results... 13118 1727204084.46018: running TaskExecutor() for managed-node2/TASK: Assert that the ansible managed comment is present in 'bond0' 13118 1727204084.46115: in run() - task 0affcd87-79f5-56a3-0a64-000000000261 13118 1727204084.46127: variable 'ansible_search_path' from source: unknown 13118 1727204084.46134: variable 'ansible_search_path' from source: unknown 13118 1727204084.46173: calling self._execute() 13118 1727204084.46276: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.46282: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.46291: variable 'omit' from source: magic vars 13118 1727204084.46689: variable 'ansible_distribution_major_version' from source: facts 13118 1727204084.46705: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204084.46712: variable 'omit' from source: magic vars 13118 1727204084.46765: variable 'omit' from source: magic vars 13118 1727204084.46883: variable 'profile' from source: include params 13118 1727204084.46887: variable 'item' from source: include params 13118 1727204084.46953: variable 'item' from source: include params 13118 1727204084.46981: variable 'omit' from source: magic vars 13118 1727204084.47087: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204084.47101: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204084.47154: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204084.47157: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204084.47160: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204084.47201: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204084.47204: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.47206: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.47388: Set connection var ansible_timeout to 10 13118 1727204084.47391: Set connection var ansible_pipelining to False 13118 1727204084.47393: Set connection var ansible_connection to ssh 13118 1727204084.47396: Set connection var ansible_shell_executable to /bin/sh 13118 1727204084.47398: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204084.47400: Set connection var ansible_shell_type to sh 13118 1727204084.47402: variable 'ansible_shell_executable' from source: unknown 13118 1727204084.47404: variable 'ansible_connection' from source: unknown 13118 1727204084.47406: variable 'ansible_module_compression' from source: unknown 13118 1727204084.47408: variable 'ansible_shell_type' from source: unknown 13118 1727204084.47410: variable 'ansible_shell_executable' from source: unknown 13118 1727204084.47412: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.47414: variable 'ansible_pipelining' from source: unknown 13118 1727204084.47416: variable 'ansible_timeout' from source: unknown 13118 1727204084.47418: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.47548: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204084.47558: variable 'omit' from source: magic vars 13118 1727204084.47565: starting attempt loop 13118 1727204084.47569: running the handler 13118 1727204084.47696: variable 'lsr_net_profile_ansible_managed' from source: set_fact 13118 1727204084.47700: Evaluated conditional (lsr_net_profile_ansible_managed): True 13118 1727204084.47708: handler run complete 13118 1727204084.47729: attempt loop complete, returning result 13118 1727204084.47735: _execute() done 13118 1727204084.47737: dumping result to json 13118 1727204084.47740: done dumping result, returning 13118 1727204084.47743: done running TaskExecutor() for managed-node2/TASK: Assert that the ansible managed comment is present in 'bond0' [0affcd87-79f5-56a3-0a64-000000000261] 13118 1727204084.47750: sending task result for task 0affcd87-79f5-56a3-0a64-000000000261 13118 1727204084.47840: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000261 13118 1727204084.47843: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false } MSG: All assertions passed 13118 1727204084.47894: no more pending results, returning what we have 13118 1727204084.47898: results queue empty 13118 1727204084.47900: checking for any_errors_fatal 13118 1727204084.47906: done checking for any_errors_fatal 13118 1727204084.47907: checking for max_fail_percentage 13118 1727204084.47909: done checking for max_fail_percentage 13118 1727204084.47910: checking to see if all hosts have failed and the running result is not ok 13118 1727204084.47911: done checking to see if all hosts have failed 13118 1727204084.47912: getting the remaining hosts for this loop 13118 1727204084.47913: done getting the remaining hosts for this loop 13118 1727204084.47917: getting the next task for host managed-node2 13118 1727204084.47924: done getting next task for host managed-node2 13118 1727204084.47927: ^ task is: TASK: Assert that the fingerprint comment is present in {{ profile }} 13118 1727204084.47930: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204084.47936: getting variables 13118 1727204084.47937: in VariableManager get_vars() 13118 1727204084.47983: Calling all_inventory to load vars for managed-node2 13118 1727204084.47986: Calling groups_inventory to load vars for managed-node2 13118 1727204084.47989: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204084.48001: Calling all_plugins_play to load vars for managed-node2 13118 1727204084.48004: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204084.48007: Calling groups_plugins_play to load vars for managed-node2 13118 1727204084.49814: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204084.51641: done with get_vars() 13118 1727204084.51683: done getting variables 13118 1727204084.51751: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204084.51875: variable 'profile' from source: include params 13118 1727204084.51879: variable 'item' from source: include params 13118 1727204084.51948: variable 'item' from source: include params TASK [Assert that the fingerprint comment is present in bond0] ***************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:15 Tuesday 24 September 2024 14:54:44 -0400 (0:00:00.066) 0:00:21.819 ***** 13118 1727204084.51990: entering _queue_task() for managed-node2/assert 13118 1727204084.52335: worker is 1 (out of 1 available) 13118 1727204084.52348: exiting _queue_task() for managed-node2/assert 13118 1727204084.52358: done queuing things up, now waiting for results queue to drain 13118 1727204084.52359: waiting for pending results... 13118 1727204084.53535: running TaskExecutor() for managed-node2/TASK: Assert that the fingerprint comment is present in bond0 13118 1727204084.53659: in run() - task 0affcd87-79f5-56a3-0a64-000000000262 13118 1727204084.53693: variable 'ansible_search_path' from source: unknown 13118 1727204084.53700: variable 'ansible_search_path' from source: unknown 13118 1727204084.53741: calling self._execute() 13118 1727204084.53848: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.53859: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.53880: variable 'omit' from source: magic vars 13118 1727204084.54580: variable 'ansible_distribution_major_version' from source: facts 13118 1727204084.54649: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204084.54750: variable 'omit' from source: magic vars 13118 1727204084.54797: variable 'omit' from source: magic vars 13118 1727204084.55025: variable 'profile' from source: include params 13118 1727204084.55037: variable 'item' from source: include params 13118 1727204084.55215: variable 'item' from source: include params 13118 1727204084.55242: variable 'omit' from source: magic vars 13118 1727204084.55304: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204084.55426: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204084.55453: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204084.55477: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204084.55621: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204084.55657: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204084.55667: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.55675: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.55786: Set connection var ansible_timeout to 10 13118 1727204084.55949: Set connection var ansible_pipelining to False 13118 1727204084.55957: Set connection var ansible_connection to ssh 13118 1727204084.55969: Set connection var ansible_shell_executable to /bin/sh 13118 1727204084.55978: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204084.55984: Set connection var ansible_shell_type to sh 13118 1727204084.56012: variable 'ansible_shell_executable' from source: unknown 13118 1727204084.56019: variable 'ansible_connection' from source: unknown 13118 1727204084.56026: variable 'ansible_module_compression' from source: unknown 13118 1727204084.56069: variable 'ansible_shell_type' from source: unknown 13118 1727204084.56077: variable 'ansible_shell_executable' from source: unknown 13118 1727204084.56083: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.56090: variable 'ansible_pipelining' from source: unknown 13118 1727204084.56097: variable 'ansible_timeout' from source: unknown 13118 1727204084.56104: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.56360: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204084.56488: variable 'omit' from source: magic vars 13118 1727204084.56497: starting attempt loop 13118 1727204084.56503: running the handler 13118 1727204084.56729: variable 'lsr_net_profile_fingerprint' from source: set_fact 13118 1727204084.56744: Evaluated conditional (lsr_net_profile_fingerprint): True 13118 1727204084.56754: handler run complete 13118 1727204084.56777: attempt loop complete, returning result 13118 1727204084.56784: _execute() done 13118 1727204084.56790: dumping result to json 13118 1727204084.56798: done dumping result, returning 13118 1727204084.56813: done running TaskExecutor() for managed-node2/TASK: Assert that the fingerprint comment is present in bond0 [0affcd87-79f5-56a3-0a64-000000000262] 13118 1727204084.56927: sending task result for task 0affcd87-79f5-56a3-0a64-000000000262 ok: [managed-node2] => { "changed": false } MSG: All assertions passed 13118 1727204084.57085: no more pending results, returning what we have 13118 1727204084.57090: results queue empty 13118 1727204084.57091: checking for any_errors_fatal 13118 1727204084.57099: done checking for any_errors_fatal 13118 1727204084.57100: checking for max_fail_percentage 13118 1727204084.57102: done checking for max_fail_percentage 13118 1727204084.57103: checking to see if all hosts have failed and the running result is not ok 13118 1727204084.57104: done checking to see if all hosts have failed 13118 1727204084.57104: getting the remaining hosts for this loop 13118 1727204084.57106: done getting the remaining hosts for this loop 13118 1727204084.57110: getting the next task for host managed-node2 13118 1727204084.57122: done getting next task for host managed-node2 13118 1727204084.57125: ^ task is: TASK: Include the task 'get_profile_stat.yml' 13118 1727204084.57128: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204084.57136: getting variables 13118 1727204084.57138: in VariableManager get_vars() 13118 1727204084.57190: Calling all_inventory to load vars for managed-node2 13118 1727204084.57193: Calling groups_inventory to load vars for managed-node2 13118 1727204084.57196: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204084.57210: Calling all_plugins_play to load vars for managed-node2 13118 1727204084.57214: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204084.57217: Calling groups_plugins_play to load vars for managed-node2 13118 1727204084.58282: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000262 13118 1727204084.58286: WORKER PROCESS EXITING 13118 1727204084.60583: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204084.62847: done with get_vars() 13118 1727204084.62879: done getting variables TASK [Include the task 'get_profile_stat.yml'] ********************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:3 Tuesday 24 September 2024 14:54:44 -0400 (0:00:00.109) 0:00:21.929 ***** 13118 1727204084.62991: entering _queue_task() for managed-node2/include_tasks 13118 1727204084.63346: worker is 1 (out of 1 available) 13118 1727204084.63359: exiting _queue_task() for managed-node2/include_tasks 13118 1727204084.63488: done queuing things up, now waiting for results queue to drain 13118 1727204084.63490: waiting for pending results... 13118 1727204084.64259: running TaskExecutor() for managed-node2/TASK: Include the task 'get_profile_stat.yml' 13118 1727204084.64773: in run() - task 0affcd87-79f5-56a3-0a64-000000000266 13118 1727204084.64777: variable 'ansible_search_path' from source: unknown 13118 1727204084.64781: variable 'ansible_search_path' from source: unknown 13118 1727204084.64784: calling self._execute() 13118 1727204084.64909: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.64918: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.64928: variable 'omit' from source: magic vars 13118 1727204084.65289: variable 'ansible_distribution_major_version' from source: facts 13118 1727204084.65300: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204084.65306: _execute() done 13118 1727204084.65311: dumping result to json 13118 1727204084.65313: done dumping result, returning 13118 1727204084.65318: done running TaskExecutor() for managed-node2/TASK: Include the task 'get_profile_stat.yml' [0affcd87-79f5-56a3-0a64-000000000266] 13118 1727204084.65325: sending task result for task 0affcd87-79f5-56a3-0a64-000000000266 13118 1727204084.65420: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000266 13118 1727204084.65425: WORKER PROCESS EXITING 13118 1727204084.65451: no more pending results, returning what we have 13118 1727204084.65457: in VariableManager get_vars() 13118 1727204084.65506: Calling all_inventory to load vars for managed-node2 13118 1727204084.65509: Calling groups_inventory to load vars for managed-node2 13118 1727204084.65511: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204084.65523: Calling all_plugins_play to load vars for managed-node2 13118 1727204084.65526: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204084.65529: Calling groups_plugins_play to load vars for managed-node2 13118 1727204084.67761: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204084.71504: done with get_vars() 13118 1727204084.71543: variable 'ansible_search_path' from source: unknown 13118 1727204084.71545: variable 'ansible_search_path' from source: unknown 13118 1727204084.71594: we have included files to process 13118 1727204084.71596: generating all_blocks data 13118 1727204084.71598: done generating all_blocks data 13118 1727204084.71604: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13118 1727204084.71606: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13118 1727204084.71610: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13118 1727204084.73055: done processing included file 13118 1727204084.73057: iterating over new_blocks loaded from include file 13118 1727204084.73059: in VariableManager get_vars() 13118 1727204084.73082: done with get_vars() 13118 1727204084.73089: filtering new block on tags 13118 1727204084.73116: done filtering new block on tags 13118 1727204084.73120: in VariableManager get_vars() 13118 1727204084.73144: done with get_vars() 13118 1727204084.73146: filtering new block on tags 13118 1727204084.73171: done filtering new block on tags 13118 1727204084.73173: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml for managed-node2 13118 1727204084.73179: extending task lists for all hosts with included blocks 13118 1727204084.73380: done extending task lists 13118 1727204084.73382: done processing included files 13118 1727204084.73383: results queue empty 13118 1727204084.73384: checking for any_errors_fatal 13118 1727204084.73387: done checking for any_errors_fatal 13118 1727204084.73388: checking for max_fail_percentage 13118 1727204084.73389: done checking for max_fail_percentage 13118 1727204084.73390: checking to see if all hosts have failed and the running result is not ok 13118 1727204084.73391: done checking to see if all hosts have failed 13118 1727204084.73392: getting the remaining hosts for this loop 13118 1727204084.73393: done getting the remaining hosts for this loop 13118 1727204084.73395: getting the next task for host managed-node2 13118 1727204084.73399: done getting next task for host managed-node2 13118 1727204084.73401: ^ task is: TASK: Initialize NM profile exist and ansible_managed comment flag 13118 1727204084.73404: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204084.73407: getting variables 13118 1727204084.73408: in VariableManager get_vars() 13118 1727204084.73425: Calling all_inventory to load vars for managed-node2 13118 1727204084.73428: Calling groups_inventory to load vars for managed-node2 13118 1727204084.73432: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204084.73438: Calling all_plugins_play to load vars for managed-node2 13118 1727204084.73441: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204084.73444: Calling groups_plugins_play to load vars for managed-node2 13118 1727204084.74843: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204084.77029: done with get_vars() 13118 1727204084.77058: done getting variables 13118 1727204084.77108: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Initialize NM profile exist and ansible_managed comment flag] ************ task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:3 Tuesday 24 September 2024 14:54:44 -0400 (0:00:00.141) 0:00:22.071 ***** 13118 1727204084.77139: entering _queue_task() for managed-node2/set_fact 13118 1727204084.77480: worker is 1 (out of 1 available) 13118 1727204084.77493: exiting _queue_task() for managed-node2/set_fact 13118 1727204084.77508: done queuing things up, now waiting for results queue to drain 13118 1727204084.77510: waiting for pending results... 13118 1727204084.78550: running TaskExecutor() for managed-node2/TASK: Initialize NM profile exist and ansible_managed comment flag 13118 1727204084.78687: in run() - task 0affcd87-79f5-56a3-0a64-0000000003f8 13118 1727204084.78887: variable 'ansible_search_path' from source: unknown 13118 1727204084.78896: variable 'ansible_search_path' from source: unknown 13118 1727204084.78939: calling self._execute() 13118 1727204084.79170: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.79182: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.79197: variable 'omit' from source: magic vars 13118 1727204084.80236: variable 'ansible_distribution_major_version' from source: facts 13118 1727204084.80255: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204084.80271: variable 'omit' from source: magic vars 13118 1727204084.80323: variable 'omit' from source: magic vars 13118 1727204084.80369: variable 'omit' from source: magic vars 13118 1727204084.80417: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204084.80459: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204084.80489: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204084.80512: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204084.80528: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204084.80565: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204084.80574: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.80581: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.80886: Set connection var ansible_timeout to 10 13118 1727204084.80902: Set connection var ansible_pipelining to False 13118 1727204084.80908: Set connection var ansible_connection to ssh 13118 1727204084.80916: Set connection var ansible_shell_executable to /bin/sh 13118 1727204084.80926: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204084.80935: Set connection var ansible_shell_type to sh 13118 1727204084.80962: variable 'ansible_shell_executable' from source: unknown 13118 1727204084.80973: variable 'ansible_connection' from source: unknown 13118 1727204084.80979: variable 'ansible_module_compression' from source: unknown 13118 1727204084.80985: variable 'ansible_shell_type' from source: unknown 13118 1727204084.80991: variable 'ansible_shell_executable' from source: unknown 13118 1727204084.80997: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.81005: variable 'ansible_pipelining' from source: unknown 13118 1727204084.81011: variable 'ansible_timeout' from source: unknown 13118 1727204084.81017: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.81157: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204084.81282: variable 'omit' from source: magic vars 13118 1727204084.81780: starting attempt loop 13118 1727204084.81788: running the handler 13118 1727204084.81805: handler run complete 13118 1727204084.81818: attempt loop complete, returning result 13118 1727204084.81824: _execute() done 13118 1727204084.81833: dumping result to json 13118 1727204084.81841: done dumping result, returning 13118 1727204084.81850: done running TaskExecutor() for managed-node2/TASK: Initialize NM profile exist and ansible_managed comment flag [0affcd87-79f5-56a3-0a64-0000000003f8] 13118 1727204084.81860: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003f8 ok: [managed-node2] => { "ansible_facts": { "lsr_net_profile_ansible_managed": false, "lsr_net_profile_exists": false, "lsr_net_profile_fingerprint": false }, "changed": false } 13118 1727204084.82018: no more pending results, returning what we have 13118 1727204084.82022: results queue empty 13118 1727204084.82023: checking for any_errors_fatal 13118 1727204084.82025: done checking for any_errors_fatal 13118 1727204084.82026: checking for max_fail_percentage 13118 1727204084.82028: done checking for max_fail_percentage 13118 1727204084.82029: checking to see if all hosts have failed and the running result is not ok 13118 1727204084.82030: done checking to see if all hosts have failed 13118 1727204084.82030: getting the remaining hosts for this loop 13118 1727204084.82032: done getting the remaining hosts for this loop 13118 1727204084.82035: getting the next task for host managed-node2 13118 1727204084.82043: done getting next task for host managed-node2 13118 1727204084.82046: ^ task is: TASK: Stat profile file 13118 1727204084.82049: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204084.82054: getting variables 13118 1727204084.82055: in VariableManager get_vars() 13118 1727204084.82100: Calling all_inventory to load vars for managed-node2 13118 1727204084.82102: Calling groups_inventory to load vars for managed-node2 13118 1727204084.82104: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204084.82118: Calling all_plugins_play to load vars for managed-node2 13118 1727204084.82121: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204084.82125: Calling groups_plugins_play to load vars for managed-node2 13118 1727204084.83374: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003f8 13118 1727204084.83379: WORKER PROCESS EXITING 13118 1727204084.84441: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204084.88999: done with get_vars() 13118 1727204084.89027: done getting variables TASK [Stat profile file] ******************************************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:9 Tuesday 24 September 2024 14:54:44 -0400 (0:00:00.121) 0:00:22.192 ***** 13118 1727204084.89249: entering _queue_task() for managed-node2/stat 13118 1727204084.89918: worker is 1 (out of 1 available) 13118 1727204084.89930: exiting _queue_task() for managed-node2/stat 13118 1727204084.89942: done queuing things up, now waiting for results queue to drain 13118 1727204084.90057: waiting for pending results... 13118 1727204084.91346: running TaskExecutor() for managed-node2/TASK: Stat profile file 13118 1727204084.91463: in run() - task 0affcd87-79f5-56a3-0a64-0000000003f9 13118 1727204084.91484: variable 'ansible_search_path' from source: unknown 13118 1727204084.91491: variable 'ansible_search_path' from source: unknown 13118 1727204084.91524: calling self._execute() 13118 1727204084.91633: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.91646: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.91661: variable 'omit' from source: magic vars 13118 1727204084.92039: variable 'ansible_distribution_major_version' from source: facts 13118 1727204084.92684: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204084.92696: variable 'omit' from source: magic vars 13118 1727204084.92755: variable 'omit' from source: magic vars 13118 1727204084.92871: variable 'profile' from source: include params 13118 1727204084.92882: variable 'item' from source: include params 13118 1727204084.92955: variable 'item' from source: include params 13118 1727204084.92982: variable 'omit' from source: magic vars 13118 1727204084.93028: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204084.93073: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204084.93100: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204084.93120: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204084.93139: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204084.93174: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204084.93182: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.93190: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.93437: Set connection var ansible_timeout to 10 13118 1727204084.93452: Set connection var ansible_pipelining to False 13118 1727204084.93457: Set connection var ansible_connection to ssh 13118 1727204084.93466: Set connection var ansible_shell_executable to /bin/sh 13118 1727204084.93474: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204084.93479: Set connection var ansible_shell_type to sh 13118 1727204084.93502: variable 'ansible_shell_executable' from source: unknown 13118 1727204084.93508: variable 'ansible_connection' from source: unknown 13118 1727204084.93516: variable 'ansible_module_compression' from source: unknown 13118 1727204084.93523: variable 'ansible_shell_type' from source: unknown 13118 1727204084.93534: variable 'ansible_shell_executable' from source: unknown 13118 1727204084.93542: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204084.93549: variable 'ansible_pipelining' from source: unknown 13118 1727204084.93556: variable 'ansible_timeout' from source: unknown 13118 1727204084.93562: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204084.93767: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204084.93784: variable 'omit' from source: magic vars 13118 1727204084.93793: starting attempt loop 13118 1727204084.93798: running the handler 13118 1727204084.93812: _low_level_execute_command(): starting 13118 1727204084.93823: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204084.94606: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204084.94623: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204084.94644: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204084.94672: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204084.94719: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204084.94734: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204084.94748: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204084.94772: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204084.94784: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204084.94794: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204084.94805: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204084.94819: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204084.94839: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204084.94853: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204084.94867: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204084.94885: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204084.94962: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204084.94982: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204084.95000: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204084.95184: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204084.96840: stdout chunk (state=3): >>>/root <<< 13118 1727204084.97038: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204084.97041: stdout chunk (state=3): >>><<< 13118 1727204084.97043: stderr chunk (state=3): >>><<< 13118 1727204084.97158: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204084.97161: _low_level_execute_command(): starting 13118 1727204084.97167: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204084.9706554-15129-130938767213503 `" && echo ansible-tmp-1727204084.9706554-15129-130938767213503="` echo /root/.ansible/tmp/ansible-tmp-1727204084.9706554-15129-130938767213503 `" ) && sleep 0' 13118 1727204084.98656: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204084.98660: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204084.98719: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204084.98723: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204084.98725: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204084.98905: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204084.98992: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204084.99145: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204085.01010: stdout chunk (state=3): >>>ansible-tmp-1727204084.9706554-15129-130938767213503=/root/.ansible/tmp/ansible-tmp-1727204084.9706554-15129-130938767213503 <<< 13118 1727204085.01125: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204085.01211: stderr chunk (state=3): >>><<< 13118 1727204085.01215: stdout chunk (state=3): >>><<< 13118 1727204085.01569: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204084.9706554-15129-130938767213503=/root/.ansible/tmp/ansible-tmp-1727204084.9706554-15129-130938767213503 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204085.01573: variable 'ansible_module_compression' from source: unknown 13118 1727204085.01576: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13118 1727204085.01578: variable 'ansible_facts' from source: unknown 13118 1727204085.01580: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204084.9706554-15129-130938767213503/AnsiballZ_stat.py 13118 1727204085.01912: Sending initial data 13118 1727204085.01915: Sent initial data (153 bytes) 13118 1727204085.04566: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204085.04585: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.04601: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.04624: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.04681: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204085.04698: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204085.04722: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.04747: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204085.04760: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204085.04775: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204085.04793: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.04809: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.04825: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.04845: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204085.04858: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204085.04875: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.04956: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204085.04981: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204085.05004: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204085.05081: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204085.06813: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204085.06850: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204085.06894: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpdwc8kruv /root/.ansible/tmp/ansible-tmp-1727204084.9706554-15129-130938767213503/AnsiballZ_stat.py <<< 13118 1727204085.06936: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204085.08052: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204085.08247: stderr chunk (state=3): >>><<< 13118 1727204085.08251: stdout chunk (state=3): >>><<< 13118 1727204085.08253: done transferring module to remote 13118 1727204085.08255: _low_level_execute_command(): starting 13118 1727204085.08257: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204084.9706554-15129-130938767213503/ /root/.ansible/tmp/ansible-tmp-1727204084.9706554-15129-130938767213503/AnsiballZ_stat.py && sleep 0' 13118 1727204085.08821: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.08824: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.08827: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.08858: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204085.08862: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.08867: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.08944: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204085.08947: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204085.08997: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204085.10801: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204085.10805: stdout chunk (state=3): >>><<< 13118 1727204085.10807: stderr chunk (state=3): >>><<< 13118 1727204085.10912: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204085.10915: _low_level_execute_command(): starting 13118 1727204085.10918: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204084.9706554-15129-130938767213503/AnsiballZ_stat.py && sleep 0' 13118 1727204085.12433: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.12438: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.12468: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.12471: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.12474: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.12558: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204085.12722: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204085.12810: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204085.25961: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/etc/sysconfig/network-scripts/ifcfg-bond0.0", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13118 1727204085.26967: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204085.27042: stderr chunk (state=3): >>><<< 13118 1727204085.27046: stdout chunk (state=3): >>><<< 13118 1727204085.27196: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/etc/sysconfig/network-scripts/ifcfg-bond0.0", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204085.27200: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/etc/sysconfig/network-scripts/ifcfg-bond0.0', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204084.9706554-15129-130938767213503/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204085.27203: _low_level_execute_command(): starting 13118 1727204085.27205: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204084.9706554-15129-130938767213503/ > /dev/null 2>&1 && sleep 0' 13118 1727204085.27855: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204085.27883: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.28672: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.28698: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.28745: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204085.28761: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204085.28789: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.28807: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204085.28819: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204085.28830: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204085.28841: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.28865: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.28904: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.28921: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204085.28931: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204085.28974: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.29052: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204085.29077: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204085.29093: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204085.29159: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204085.31077: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204085.31081: stdout chunk (state=3): >>><<< 13118 1727204085.31084: stderr chunk (state=3): >>><<< 13118 1727204085.31374: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204085.31378: handler run complete 13118 1727204085.31380: attempt loop complete, returning result 13118 1727204085.31382: _execute() done 13118 1727204085.31385: dumping result to json 13118 1727204085.31387: done dumping result, returning 13118 1727204085.31389: done running TaskExecutor() for managed-node2/TASK: Stat profile file [0affcd87-79f5-56a3-0a64-0000000003f9] 13118 1727204085.31391: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003f9 13118 1727204085.31468: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003f9 13118 1727204085.31471: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "stat": { "exists": false } } 13118 1727204085.31551: no more pending results, returning what we have 13118 1727204085.31557: results queue empty 13118 1727204085.31558: checking for any_errors_fatal 13118 1727204085.31571: done checking for any_errors_fatal 13118 1727204085.31572: checking for max_fail_percentage 13118 1727204085.31574: done checking for max_fail_percentage 13118 1727204085.31575: checking to see if all hosts have failed and the running result is not ok 13118 1727204085.31576: done checking to see if all hosts have failed 13118 1727204085.31576: getting the remaining hosts for this loop 13118 1727204085.31578: done getting the remaining hosts for this loop 13118 1727204085.31582: getting the next task for host managed-node2 13118 1727204085.31591: done getting next task for host managed-node2 13118 1727204085.31594: ^ task is: TASK: Set NM profile exist flag based on the profile files 13118 1727204085.31599: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204085.31604: getting variables 13118 1727204085.31605: in VariableManager get_vars() 13118 1727204085.31655: Calling all_inventory to load vars for managed-node2 13118 1727204085.31658: Calling groups_inventory to load vars for managed-node2 13118 1727204085.31661: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204085.31675: Calling all_plugins_play to load vars for managed-node2 13118 1727204085.31678: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204085.31681: Calling groups_plugins_play to load vars for managed-node2 13118 1727204085.33441: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204085.35309: done with get_vars() 13118 1727204085.35342: done getting variables 13118 1727204085.35415: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Set NM profile exist flag based on the profile files] ******************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:17 Tuesday 24 September 2024 14:54:45 -0400 (0:00:00.461) 0:00:22.654 ***** 13118 1727204085.35453: entering _queue_task() for managed-node2/set_fact 13118 1727204085.35899: worker is 1 (out of 1 available) 13118 1727204085.35926: exiting _queue_task() for managed-node2/set_fact 13118 1727204085.35950: done queuing things up, now waiting for results queue to drain 13118 1727204085.35951: waiting for pending results... 13118 1727204085.36295: running TaskExecutor() for managed-node2/TASK: Set NM profile exist flag based on the profile files 13118 1727204085.36482: in run() - task 0affcd87-79f5-56a3-0a64-0000000003fa 13118 1727204085.36502: variable 'ansible_search_path' from source: unknown 13118 1727204085.36510: variable 'ansible_search_path' from source: unknown 13118 1727204085.36558: calling self._execute() 13118 1727204085.36690: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204085.36703: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204085.36717: variable 'omit' from source: magic vars 13118 1727204085.37241: variable 'ansible_distribution_major_version' from source: facts 13118 1727204085.37260: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204085.37442: variable 'profile_stat' from source: set_fact 13118 1727204085.37465: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204085.37474: when evaluation is False, skipping this task 13118 1727204085.37482: _execute() done 13118 1727204085.37490: dumping result to json 13118 1727204085.37503: done dumping result, returning 13118 1727204085.37525: done running TaskExecutor() for managed-node2/TASK: Set NM profile exist flag based on the profile files [0affcd87-79f5-56a3-0a64-0000000003fa] 13118 1727204085.37547: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003fa skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204085.37703: no more pending results, returning what we have 13118 1727204085.37709: results queue empty 13118 1727204085.37710: checking for any_errors_fatal 13118 1727204085.37719: done checking for any_errors_fatal 13118 1727204085.37720: checking for max_fail_percentage 13118 1727204085.37723: done checking for max_fail_percentage 13118 1727204085.37724: checking to see if all hosts have failed and the running result is not ok 13118 1727204085.37725: done checking to see if all hosts have failed 13118 1727204085.37725: getting the remaining hosts for this loop 13118 1727204085.37727: done getting the remaining hosts for this loop 13118 1727204085.37734: getting the next task for host managed-node2 13118 1727204085.37743: done getting next task for host managed-node2 13118 1727204085.37747: ^ task is: TASK: Get NM profile info 13118 1727204085.37751: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204085.37756: getting variables 13118 1727204085.37758: in VariableManager get_vars() 13118 1727204085.37806: Calling all_inventory to load vars for managed-node2 13118 1727204085.37809: Calling groups_inventory to load vars for managed-node2 13118 1727204085.37812: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204085.37826: Calling all_plugins_play to load vars for managed-node2 13118 1727204085.37829: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204085.37836: Calling groups_plugins_play to load vars for managed-node2 13118 1727204085.38866: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003fa 13118 1727204085.38871: WORKER PROCESS EXITING 13118 1727204085.44809: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204085.46406: done with get_vars() 13118 1727204085.46438: done getting variables 13118 1727204085.46492: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Get NM profile info] ***************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:25 Tuesday 24 September 2024 14:54:45 -0400 (0:00:00.110) 0:00:22.765 ***** 13118 1727204085.46523: entering _queue_task() for managed-node2/shell 13118 1727204085.46863: worker is 1 (out of 1 available) 13118 1727204085.46879: exiting _queue_task() for managed-node2/shell 13118 1727204085.46892: done queuing things up, now waiting for results queue to drain 13118 1727204085.46894: waiting for pending results... 13118 1727204085.47233: running TaskExecutor() for managed-node2/TASK: Get NM profile info 13118 1727204085.47350: in run() - task 0affcd87-79f5-56a3-0a64-0000000003fb 13118 1727204085.47361: variable 'ansible_search_path' from source: unknown 13118 1727204085.47366: variable 'ansible_search_path' from source: unknown 13118 1727204085.47404: calling self._execute() 13118 1727204085.47568: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204085.47572: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204085.47577: variable 'omit' from source: magic vars 13118 1727204085.47906: variable 'ansible_distribution_major_version' from source: facts 13118 1727204085.47923: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204085.47933: variable 'omit' from source: magic vars 13118 1727204085.47979: variable 'omit' from source: magic vars 13118 1727204085.48086: variable 'profile' from source: include params 13118 1727204085.48091: variable 'item' from source: include params 13118 1727204085.48451: variable 'item' from source: include params 13118 1727204085.48471: variable 'omit' from source: magic vars 13118 1727204085.48517: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204085.48549: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204085.48571: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204085.48586: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204085.48597: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204085.48633: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204085.48637: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204085.48639: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204085.48736: Set connection var ansible_timeout to 10 13118 1727204085.48749: Set connection var ansible_pipelining to False 13118 1727204085.48752: Set connection var ansible_connection to ssh 13118 1727204085.48757: Set connection var ansible_shell_executable to /bin/sh 13118 1727204085.48765: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204085.48769: Set connection var ansible_shell_type to sh 13118 1727204085.48791: variable 'ansible_shell_executable' from source: unknown 13118 1727204085.48795: variable 'ansible_connection' from source: unknown 13118 1727204085.48798: variable 'ansible_module_compression' from source: unknown 13118 1727204085.48800: variable 'ansible_shell_type' from source: unknown 13118 1727204085.48803: variable 'ansible_shell_executable' from source: unknown 13118 1727204085.48805: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204085.48808: variable 'ansible_pipelining' from source: unknown 13118 1727204085.48811: variable 'ansible_timeout' from source: unknown 13118 1727204085.48813: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204085.48942: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204085.48958: variable 'omit' from source: magic vars 13118 1727204085.48965: starting attempt loop 13118 1727204085.48968: running the handler 13118 1727204085.48979: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204085.48996: _low_level_execute_command(): starting 13118 1727204085.49005: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204085.50234: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204085.50338: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.50348: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.50362: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.50409: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204085.50490: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204085.50500: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.50513: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204085.50553: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204085.50559: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204085.50570: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.50581: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.50612: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.50637: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204085.50645: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204085.50678: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.50941: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204085.50957: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204085.50960: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204085.51076: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204085.52706: stdout chunk (state=3): >>>/root <<< 13118 1727204085.52999: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204085.53399: stderr chunk (state=3): >>><<< 13118 1727204085.53402: stdout chunk (state=3): >>><<< 13118 1727204085.53538: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204085.53550: _low_level_execute_command(): starting 13118 1727204085.53554: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204085.5342448-15151-128842638639557 `" && echo ansible-tmp-1727204085.5342448-15151-128842638639557="` echo /root/.ansible/tmp/ansible-tmp-1727204085.5342448-15151-128842638639557 `" ) && sleep 0' 13118 1727204085.54657: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.54661: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.54693: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.54696: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.54698: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204085.54701: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.54770: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204085.54777: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204085.55837: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204085.55851: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204085.57722: stdout chunk (state=3): >>>ansible-tmp-1727204085.5342448-15151-128842638639557=/root/.ansible/tmp/ansible-tmp-1727204085.5342448-15151-128842638639557 <<< 13118 1727204085.57822: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204085.57903: stderr chunk (state=3): >>><<< 13118 1727204085.57907: stdout chunk (state=3): >>><<< 13118 1727204085.57970: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204085.5342448-15151-128842638639557=/root/.ansible/tmp/ansible-tmp-1727204085.5342448-15151-128842638639557 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204085.57973: variable 'ansible_module_compression' from source: unknown 13118 1727204085.58076: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13118 1727204085.58079: variable 'ansible_facts' from source: unknown 13118 1727204085.58148: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204085.5342448-15151-128842638639557/AnsiballZ_command.py 13118 1727204085.58881: Sending initial data 13118 1727204085.58884: Sent initial data (156 bytes) 13118 1727204085.61373: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.61377: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.61522: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.61526: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.61528: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.61592: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204085.61704: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204085.61857: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204085.63610: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204085.63658: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204085.63695: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpk3hegnhz /root/.ansible/tmp/ansible-tmp-1727204085.5342448-15151-128842638639557/AnsiballZ_command.py <<< 13118 1727204085.63745: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204085.65171: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204085.65209: stderr chunk (state=3): >>><<< 13118 1727204085.65212: stdout chunk (state=3): >>><<< 13118 1727204085.65238: done transferring module to remote 13118 1727204085.65250: _low_level_execute_command(): starting 13118 1727204085.65253: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204085.5342448-15151-128842638639557/ /root/.ansible/tmp/ansible-tmp-1727204085.5342448-15151-128842638639557/AnsiballZ_command.py && sleep 0' 13118 1727204085.68066: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204085.68076: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.68085: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.68099: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.68140: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204085.68148: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204085.68155: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.68170: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204085.68178: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204085.68186: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204085.68193: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.68201: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.68213: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.68220: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204085.68226: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204085.68236: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.68309: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204085.68324: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204085.68327: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204085.68931: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204085.70768: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204085.70773: stdout chunk (state=3): >>><<< 13118 1727204085.70778: stderr chunk (state=3): >>><<< 13118 1727204085.70797: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204085.70800: _low_level_execute_command(): starting 13118 1727204085.70805: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204085.5342448-15151-128842638639557/AnsiballZ_command.py && sleep 0' 13118 1727204085.72535: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204085.72657: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.72669: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.72684: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.72825: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204085.72832: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204085.72835: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.72837: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204085.72840: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204085.72842: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204085.72844: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.72846: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.72849: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.72851: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204085.72853: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204085.72855: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.72936: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204085.73041: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204085.73051: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204085.73184: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204085.88704: stdout chunk (state=3): >>> {"changed": true, "stdout": "bond0.0 /etc/NetworkManager/system-connections/bond0.0.nmconnection ", "stderr": "", "rc": 0, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0.0 | grep /etc", "start": "2024-09-24 14:54:45.862846", "end": "2024-09-24 14:54:45.886208", "delta": "0:00:00.023362", "msg": "", "invocation": {"module_args": {"_raw_params": "nmcli -f NAME,FILENAME connection show |grep bond0.0 | grep /etc", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13118 1727204085.89971: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204085.89976: stdout chunk (state=3): >>><<< 13118 1727204085.89979: stderr chunk (state=3): >>><<< 13118 1727204085.90006: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "bond0.0 /etc/NetworkManager/system-connections/bond0.0.nmconnection ", "stderr": "", "rc": 0, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0.0 | grep /etc", "start": "2024-09-24 14:54:45.862846", "end": "2024-09-24 14:54:45.886208", "delta": "0:00:00.023362", "msg": "", "invocation": {"module_args": {"_raw_params": "nmcli -f NAME,FILENAME connection show |grep bond0.0 | grep /etc", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204085.90045: done with _execute_module (ansible.legacy.command, {'_raw_params': 'nmcli -f NAME,FILENAME connection show |grep bond0.0 | grep /etc', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204085.5342448-15151-128842638639557/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204085.90054: _low_level_execute_command(): starting 13118 1727204085.90059: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204085.5342448-15151-128842638639557/ > /dev/null 2>&1 && sleep 0' 13118 1727204085.90960: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204085.91484: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.91503: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.91524: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.91571: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204085.91584: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204085.91596: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.91612: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204085.91622: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204085.91630: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204085.91640: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204085.91651: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204085.91665: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204085.91677: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204085.91688: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204085.91702: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204085.91785: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204085.91809: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204085.91827: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204085.91903: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204085.93801: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204085.93806: stdout chunk (state=3): >>><<< 13118 1727204085.93808: stderr chunk (state=3): >>><<< 13118 1727204085.93873: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204085.93878: handler run complete 13118 1727204085.93976: Evaluated conditional (False): False 13118 1727204085.93979: attempt loop complete, returning result 13118 1727204085.93981: _execute() done 13118 1727204085.93983: dumping result to json 13118 1727204085.93984: done dumping result, returning 13118 1727204085.93986: done running TaskExecutor() for managed-node2/TASK: Get NM profile info [0affcd87-79f5-56a3-0a64-0000000003fb] 13118 1727204085.93988: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003fb 13118 1727204085.94060: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003fb 13118 1727204085.94063: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0.0 | grep /etc", "delta": "0:00:00.023362", "end": "2024-09-24 14:54:45.886208", "rc": 0, "start": "2024-09-24 14:54:45.862846" } STDOUT: bond0.0 /etc/NetworkManager/system-connections/bond0.0.nmconnection 13118 1727204085.94142: no more pending results, returning what we have 13118 1727204085.94146: results queue empty 13118 1727204085.94147: checking for any_errors_fatal 13118 1727204085.94154: done checking for any_errors_fatal 13118 1727204085.94155: checking for max_fail_percentage 13118 1727204085.94157: done checking for max_fail_percentage 13118 1727204085.94157: checking to see if all hosts have failed and the running result is not ok 13118 1727204085.94158: done checking to see if all hosts have failed 13118 1727204085.94159: getting the remaining hosts for this loop 13118 1727204085.94160: done getting the remaining hosts for this loop 13118 1727204085.94166: getting the next task for host managed-node2 13118 1727204085.94174: done getting next task for host managed-node2 13118 1727204085.94177: ^ task is: TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output 13118 1727204085.94181: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204085.94188: getting variables 13118 1727204085.94189: in VariableManager get_vars() 13118 1727204085.94229: Calling all_inventory to load vars for managed-node2 13118 1727204085.94232: Calling groups_inventory to load vars for managed-node2 13118 1727204085.94234: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204085.94246: Calling all_plugins_play to load vars for managed-node2 13118 1727204085.94248: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204085.94250: Calling groups_plugins_play to load vars for managed-node2 13118 1727204085.95879: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204085.97886: done with get_vars() 13118 1727204085.97916: done getting variables 13118 1727204085.97990: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Set NM profile exist flag and ansible_managed flag true based on the nmcli output] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:35 Tuesday 24 September 2024 14:54:45 -0400 (0:00:00.514) 0:00:23.280 ***** 13118 1727204085.98028: entering _queue_task() for managed-node2/set_fact 13118 1727204085.98386: worker is 1 (out of 1 available) 13118 1727204085.98400: exiting _queue_task() for managed-node2/set_fact 13118 1727204085.98414: done queuing things up, now waiting for results queue to drain 13118 1727204085.98416: waiting for pending results... 13118 1727204085.98734: running TaskExecutor() for managed-node2/TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output 13118 1727204085.98883: in run() - task 0affcd87-79f5-56a3-0a64-0000000003fc 13118 1727204085.98903: variable 'ansible_search_path' from source: unknown 13118 1727204085.98914: variable 'ansible_search_path' from source: unknown 13118 1727204085.98955: calling self._execute() 13118 1727204085.99068: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204085.99081: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204085.99101: variable 'omit' from source: magic vars 13118 1727204085.99532: variable 'ansible_distribution_major_version' from source: facts 13118 1727204085.99552: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204085.99712: variable 'nm_profile_exists' from source: set_fact 13118 1727204085.99740: Evaluated conditional (nm_profile_exists.rc == 0): True 13118 1727204085.99758: variable 'omit' from source: magic vars 13118 1727204085.99815: variable 'omit' from source: magic vars 13118 1727204085.99867: variable 'omit' from source: magic vars 13118 1727204085.99943: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204085.99995: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204086.00350: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204086.00376: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204086.00396: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204086.00432: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204086.00441: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.00448: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.00562: Set connection var ansible_timeout to 10 13118 1727204086.00583: Set connection var ansible_pipelining to False 13118 1727204086.00589: Set connection var ansible_connection to ssh 13118 1727204086.00602: Set connection var ansible_shell_executable to /bin/sh 13118 1727204086.00611: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204086.00617: Set connection var ansible_shell_type to sh 13118 1727204086.00647: variable 'ansible_shell_executable' from source: unknown 13118 1727204086.00655: variable 'ansible_connection' from source: unknown 13118 1727204086.00661: variable 'ansible_module_compression' from source: unknown 13118 1727204086.00672: variable 'ansible_shell_type' from source: unknown 13118 1727204086.00679: variable 'ansible_shell_executable' from source: unknown 13118 1727204086.00684: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.00692: variable 'ansible_pipelining' from source: unknown 13118 1727204086.00698: variable 'ansible_timeout' from source: unknown 13118 1727204086.00709: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.00859: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204086.00940: variable 'omit' from source: magic vars 13118 1727204086.00978: starting attempt loop 13118 1727204086.00985: running the handler 13118 1727204086.01049: handler run complete 13118 1727204086.01066: attempt loop complete, returning result 13118 1727204086.01074: _execute() done 13118 1727204086.01080: dumping result to json 13118 1727204086.01087: done dumping result, returning 13118 1727204086.01149: done running TaskExecutor() for managed-node2/TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output [0affcd87-79f5-56a3-0a64-0000000003fc] 13118 1727204086.01160: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003fc ok: [managed-node2] => { "ansible_facts": { "lsr_net_profile_ansible_managed": true, "lsr_net_profile_exists": true, "lsr_net_profile_fingerprint": true }, "changed": false } 13118 1727204086.01318: no more pending results, returning what we have 13118 1727204086.01323: results queue empty 13118 1727204086.01324: checking for any_errors_fatal 13118 1727204086.01336: done checking for any_errors_fatal 13118 1727204086.01337: checking for max_fail_percentage 13118 1727204086.01338: done checking for max_fail_percentage 13118 1727204086.01340: checking to see if all hosts have failed and the running result is not ok 13118 1727204086.01340: done checking to see if all hosts have failed 13118 1727204086.01341: getting the remaining hosts for this loop 13118 1727204086.01343: done getting the remaining hosts for this loop 13118 1727204086.01346: getting the next task for host managed-node2 13118 1727204086.01357: done getting next task for host managed-node2 13118 1727204086.01360: ^ task is: TASK: Get the ansible_managed comment in ifcfg-{{ profile }} 13118 1727204086.01365: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204086.01370: getting variables 13118 1727204086.01371: in VariableManager get_vars() 13118 1727204086.01414: Calling all_inventory to load vars for managed-node2 13118 1727204086.01416: Calling groups_inventory to load vars for managed-node2 13118 1727204086.01419: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204086.01432: Calling all_plugins_play to load vars for managed-node2 13118 1727204086.01435: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204086.01438: Calling groups_plugins_play to load vars for managed-node2 13118 1727204086.02486: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003fc 13118 1727204086.02491: WORKER PROCESS EXITING 13118 1727204086.04196: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204086.06115: done with get_vars() 13118 1727204086.06157: done getting variables 13118 1727204086.06245: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204086.06386: variable 'profile' from source: include params 13118 1727204086.06390: variable 'item' from source: include params 13118 1727204086.06455: variable 'item' from source: include params TASK [Get the ansible_managed comment in ifcfg-bond0.0] ************************ task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:49 Tuesday 24 September 2024 14:54:46 -0400 (0:00:00.084) 0:00:23.364 ***** 13118 1727204086.06496: entering _queue_task() for managed-node2/command 13118 1727204086.06942: worker is 1 (out of 1 available) 13118 1727204086.06958: exiting _queue_task() for managed-node2/command 13118 1727204086.06974: done queuing things up, now waiting for results queue to drain 13118 1727204086.06975: waiting for pending results... 13118 1727204086.07222: running TaskExecutor() for managed-node2/TASK: Get the ansible_managed comment in ifcfg-bond0.0 13118 1727204086.07316: in run() - task 0affcd87-79f5-56a3-0a64-0000000003fe 13118 1727204086.07327: variable 'ansible_search_path' from source: unknown 13118 1727204086.07331: variable 'ansible_search_path' from source: unknown 13118 1727204086.07361: calling self._execute() 13118 1727204086.07437: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.07443: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.07451: variable 'omit' from source: magic vars 13118 1727204086.07720: variable 'ansible_distribution_major_version' from source: facts 13118 1727204086.07729: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204086.07833: variable 'profile_stat' from source: set_fact 13118 1727204086.07848: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204086.07851: when evaluation is False, skipping this task 13118 1727204086.07854: _execute() done 13118 1727204086.07857: dumping result to json 13118 1727204086.07859: done dumping result, returning 13118 1727204086.08152: done running TaskExecutor() for managed-node2/TASK: Get the ansible_managed comment in ifcfg-bond0.0 [0affcd87-79f5-56a3-0a64-0000000003fe] 13118 1727204086.08155: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003fe 13118 1727204086.08224: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003fe 13118 1727204086.08228: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204086.08295: no more pending results, returning what we have 13118 1727204086.08299: results queue empty 13118 1727204086.08300: checking for any_errors_fatal 13118 1727204086.08305: done checking for any_errors_fatal 13118 1727204086.08305: checking for max_fail_percentage 13118 1727204086.08307: done checking for max_fail_percentage 13118 1727204086.08308: checking to see if all hosts have failed and the running result is not ok 13118 1727204086.08309: done checking to see if all hosts have failed 13118 1727204086.08310: getting the remaining hosts for this loop 13118 1727204086.08311: done getting the remaining hosts for this loop 13118 1727204086.08314: getting the next task for host managed-node2 13118 1727204086.08321: done getting next task for host managed-node2 13118 1727204086.08323: ^ task is: TASK: Verify the ansible_managed comment in ifcfg-{{ profile }} 13118 1727204086.08327: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204086.08332: getting variables 13118 1727204086.08333: in VariableManager get_vars() 13118 1727204086.08380: Calling all_inventory to load vars for managed-node2 13118 1727204086.08383: Calling groups_inventory to load vars for managed-node2 13118 1727204086.08386: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204086.08395: Calling all_plugins_play to load vars for managed-node2 13118 1727204086.08398: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204086.08400: Calling groups_plugins_play to load vars for managed-node2 13118 1727204086.09578: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204086.10519: done with get_vars() 13118 1727204086.10544: done getting variables 13118 1727204086.10591: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204086.10684: variable 'profile' from source: include params 13118 1727204086.10687: variable 'item' from source: include params 13118 1727204086.10727: variable 'item' from source: include params TASK [Verify the ansible_managed comment in ifcfg-bond0.0] ********************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:56 Tuesday 24 September 2024 14:54:46 -0400 (0:00:00.042) 0:00:23.407 ***** 13118 1727204086.10755: entering _queue_task() for managed-node2/set_fact 13118 1727204086.11024: worker is 1 (out of 1 available) 13118 1727204086.11173: exiting _queue_task() for managed-node2/set_fact 13118 1727204086.11196: done queuing things up, now waiting for results queue to drain 13118 1727204086.11210: waiting for pending results... 13118 1727204086.11337: running TaskExecutor() for managed-node2/TASK: Verify the ansible_managed comment in ifcfg-bond0.0 13118 1727204086.11495: in run() - task 0affcd87-79f5-56a3-0a64-0000000003ff 13118 1727204086.11521: variable 'ansible_search_path' from source: unknown 13118 1727204086.11536: variable 'ansible_search_path' from source: unknown 13118 1727204086.11581: calling self._execute() 13118 1727204086.11708: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.11719: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.11737: variable 'omit' from source: magic vars 13118 1727204086.12150: variable 'ansible_distribution_major_version' from source: facts 13118 1727204086.12173: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204086.12321: variable 'profile_stat' from source: set_fact 13118 1727204086.12345: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204086.12356: when evaluation is False, skipping this task 13118 1727204086.12365: _execute() done 13118 1727204086.12372: dumping result to json 13118 1727204086.12379: done dumping result, returning 13118 1727204086.12388: done running TaskExecutor() for managed-node2/TASK: Verify the ansible_managed comment in ifcfg-bond0.0 [0affcd87-79f5-56a3-0a64-0000000003ff] 13118 1727204086.12406: sending task result for task 0affcd87-79f5-56a3-0a64-0000000003ff skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204086.12568: no more pending results, returning what we have 13118 1727204086.12573: results queue empty 13118 1727204086.12575: checking for any_errors_fatal 13118 1727204086.12584: done checking for any_errors_fatal 13118 1727204086.12585: checking for max_fail_percentage 13118 1727204086.12587: done checking for max_fail_percentage 13118 1727204086.12588: checking to see if all hosts have failed and the running result is not ok 13118 1727204086.12589: done checking to see if all hosts have failed 13118 1727204086.12589: getting the remaining hosts for this loop 13118 1727204086.12591: done getting the remaining hosts for this loop 13118 1727204086.12595: getting the next task for host managed-node2 13118 1727204086.12602: done getting next task for host managed-node2 13118 1727204086.12605: ^ task is: TASK: Get the fingerprint comment in ifcfg-{{ profile }} 13118 1727204086.12610: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204086.12616: getting variables 13118 1727204086.12618: in VariableManager get_vars() 13118 1727204086.12695: Calling all_inventory to load vars for managed-node2 13118 1727204086.12699: Calling groups_inventory to load vars for managed-node2 13118 1727204086.12702: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204086.12716: Calling all_plugins_play to load vars for managed-node2 13118 1727204086.12720: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204086.12724: Calling groups_plugins_play to load vars for managed-node2 13118 1727204086.13399: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000003ff 13118 1727204086.13403: WORKER PROCESS EXITING 13118 1727204086.13916: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204086.15111: done with get_vars() 13118 1727204086.15145: done getting variables 13118 1727204086.15212: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204086.15348: variable 'profile' from source: include params 13118 1727204086.15352: variable 'item' from source: include params 13118 1727204086.15424: variable 'item' from source: include params TASK [Get the fingerprint comment in ifcfg-bond0.0] **************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:62 Tuesday 24 September 2024 14:54:46 -0400 (0:00:00.047) 0:00:23.454 ***** 13118 1727204086.15462: entering _queue_task() for managed-node2/command 13118 1727204086.15786: worker is 1 (out of 1 available) 13118 1727204086.15797: exiting _queue_task() for managed-node2/command 13118 1727204086.15808: done queuing things up, now waiting for results queue to drain 13118 1727204086.15810: waiting for pending results... 13118 1727204086.16146: running TaskExecutor() for managed-node2/TASK: Get the fingerprint comment in ifcfg-bond0.0 13118 1727204086.16265: in run() - task 0affcd87-79f5-56a3-0a64-000000000400 13118 1727204086.16286: variable 'ansible_search_path' from source: unknown 13118 1727204086.16291: variable 'ansible_search_path' from source: unknown 13118 1727204086.16326: calling self._execute() 13118 1727204086.16440: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.16444: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.16454: variable 'omit' from source: magic vars 13118 1727204086.16854: variable 'ansible_distribution_major_version' from source: facts 13118 1727204086.16873: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204086.17088: variable 'profile_stat' from source: set_fact 13118 1727204086.17106: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204086.17115: when evaluation is False, skipping this task 13118 1727204086.17118: _execute() done 13118 1727204086.17121: dumping result to json 13118 1727204086.17124: done dumping result, returning 13118 1727204086.17132: done running TaskExecutor() for managed-node2/TASK: Get the fingerprint comment in ifcfg-bond0.0 [0affcd87-79f5-56a3-0a64-000000000400] 13118 1727204086.17152: sending task result for task 0affcd87-79f5-56a3-0a64-000000000400 skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204086.17311: no more pending results, returning what we have 13118 1727204086.17316: results queue empty 13118 1727204086.17317: checking for any_errors_fatal 13118 1727204086.17323: done checking for any_errors_fatal 13118 1727204086.17324: checking for max_fail_percentage 13118 1727204086.17325: done checking for max_fail_percentage 13118 1727204086.17326: checking to see if all hosts have failed and the running result is not ok 13118 1727204086.17327: done checking to see if all hosts have failed 13118 1727204086.17328: getting the remaining hosts for this loop 13118 1727204086.17329: done getting the remaining hosts for this loop 13118 1727204086.17335: getting the next task for host managed-node2 13118 1727204086.17342: done getting next task for host managed-node2 13118 1727204086.17345: ^ task is: TASK: Verify the fingerprint comment in ifcfg-{{ profile }} 13118 1727204086.17349: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204086.17353: getting variables 13118 1727204086.17355: in VariableManager get_vars() 13118 1727204086.17398: Calling all_inventory to load vars for managed-node2 13118 1727204086.17401: Calling groups_inventory to load vars for managed-node2 13118 1727204086.17403: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204086.17416: Calling all_plugins_play to load vars for managed-node2 13118 1727204086.17418: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204086.17422: Calling groups_plugins_play to load vars for managed-node2 13118 1727204086.17949: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000400 13118 1727204086.17954: WORKER PROCESS EXITING 13118 1727204086.18296: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204086.19665: done with get_vars() 13118 1727204086.19688: done getting variables 13118 1727204086.19755: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204086.19881: variable 'profile' from source: include params 13118 1727204086.19885: variable 'item' from source: include params 13118 1727204086.19949: variable 'item' from source: include params TASK [Verify the fingerprint comment in ifcfg-bond0.0] ************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:69 Tuesday 24 September 2024 14:54:46 -0400 (0:00:00.045) 0:00:23.499 ***** 13118 1727204086.19984: entering _queue_task() for managed-node2/set_fact 13118 1727204086.20303: worker is 1 (out of 1 available) 13118 1727204086.20318: exiting _queue_task() for managed-node2/set_fact 13118 1727204086.20329: done queuing things up, now waiting for results queue to drain 13118 1727204086.20333: waiting for pending results... 13118 1727204086.20517: running TaskExecutor() for managed-node2/TASK: Verify the fingerprint comment in ifcfg-bond0.0 13118 1727204086.20596: in run() - task 0affcd87-79f5-56a3-0a64-000000000401 13118 1727204086.20608: variable 'ansible_search_path' from source: unknown 13118 1727204086.20611: variable 'ansible_search_path' from source: unknown 13118 1727204086.20642: calling self._execute() 13118 1727204086.20720: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.20725: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.20736: variable 'omit' from source: magic vars 13118 1727204086.21011: variable 'ansible_distribution_major_version' from source: facts 13118 1727204086.21021: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204086.21111: variable 'profile_stat' from source: set_fact 13118 1727204086.21119: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204086.21122: when evaluation is False, skipping this task 13118 1727204086.21124: _execute() done 13118 1727204086.21127: dumping result to json 13118 1727204086.21131: done dumping result, returning 13118 1727204086.21139: done running TaskExecutor() for managed-node2/TASK: Verify the fingerprint comment in ifcfg-bond0.0 [0affcd87-79f5-56a3-0a64-000000000401] 13118 1727204086.21145: sending task result for task 0affcd87-79f5-56a3-0a64-000000000401 13118 1727204086.21231: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000401 13118 1727204086.21234: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204086.21284: no more pending results, returning what we have 13118 1727204086.21288: results queue empty 13118 1727204086.21289: checking for any_errors_fatal 13118 1727204086.21296: done checking for any_errors_fatal 13118 1727204086.21297: checking for max_fail_percentage 13118 1727204086.21298: done checking for max_fail_percentage 13118 1727204086.21299: checking to see if all hosts have failed and the running result is not ok 13118 1727204086.21300: done checking to see if all hosts have failed 13118 1727204086.21301: getting the remaining hosts for this loop 13118 1727204086.21302: done getting the remaining hosts for this loop 13118 1727204086.21306: getting the next task for host managed-node2 13118 1727204086.21315: done getting next task for host managed-node2 13118 1727204086.21318: ^ task is: TASK: Assert that the profile is present - '{{ profile }}' 13118 1727204086.21321: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204086.21325: getting variables 13118 1727204086.21326: in VariableManager get_vars() 13118 1727204086.21373: Calling all_inventory to load vars for managed-node2 13118 1727204086.21376: Calling groups_inventory to load vars for managed-node2 13118 1727204086.21378: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204086.21388: Calling all_plugins_play to load vars for managed-node2 13118 1727204086.21390: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204086.21393: Calling groups_plugins_play to load vars for managed-node2 13118 1727204086.22222: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204086.23171: done with get_vars() 13118 1727204086.23191: done getting variables 13118 1727204086.23239: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204086.23333: variable 'profile' from source: include params 13118 1727204086.23336: variable 'item' from source: include params 13118 1727204086.23379: variable 'item' from source: include params TASK [Assert that the profile is present - 'bond0.0'] ************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:5 Tuesday 24 September 2024 14:54:46 -0400 (0:00:00.034) 0:00:23.534 ***** 13118 1727204086.23405: entering _queue_task() for managed-node2/assert 13118 1727204086.23649: worker is 1 (out of 1 available) 13118 1727204086.23663: exiting _queue_task() for managed-node2/assert 13118 1727204086.23678: done queuing things up, now waiting for results queue to drain 13118 1727204086.23680: waiting for pending results... 13118 1727204086.23866: running TaskExecutor() for managed-node2/TASK: Assert that the profile is present - 'bond0.0' 13118 1727204086.23936: in run() - task 0affcd87-79f5-56a3-0a64-000000000267 13118 1727204086.23945: variable 'ansible_search_path' from source: unknown 13118 1727204086.23950: variable 'ansible_search_path' from source: unknown 13118 1727204086.23981: calling self._execute() 13118 1727204086.24060: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.24063: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.24074: variable 'omit' from source: magic vars 13118 1727204086.24339: variable 'ansible_distribution_major_version' from source: facts 13118 1727204086.24350: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204086.24356: variable 'omit' from source: magic vars 13118 1727204086.24395: variable 'omit' from source: magic vars 13118 1727204086.24468: variable 'profile' from source: include params 13118 1727204086.24474: variable 'item' from source: include params 13118 1727204086.24521: variable 'item' from source: include params 13118 1727204086.24536: variable 'omit' from source: magic vars 13118 1727204086.24573: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204086.24601: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204086.24618: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204086.24634: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204086.24643: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204086.24667: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204086.24671: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.24673: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.24746: Set connection var ansible_timeout to 10 13118 1727204086.24756: Set connection var ansible_pipelining to False 13118 1727204086.24758: Set connection var ansible_connection to ssh 13118 1727204086.24765: Set connection var ansible_shell_executable to /bin/sh 13118 1727204086.24770: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204086.24773: Set connection var ansible_shell_type to sh 13118 1727204086.24791: variable 'ansible_shell_executable' from source: unknown 13118 1727204086.24794: variable 'ansible_connection' from source: unknown 13118 1727204086.24796: variable 'ansible_module_compression' from source: unknown 13118 1727204086.24799: variable 'ansible_shell_type' from source: unknown 13118 1727204086.24801: variable 'ansible_shell_executable' from source: unknown 13118 1727204086.24803: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.24807: variable 'ansible_pipelining' from source: unknown 13118 1727204086.24812: variable 'ansible_timeout' from source: unknown 13118 1727204086.24814: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.24916: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204086.24928: variable 'omit' from source: magic vars 13118 1727204086.24934: starting attempt loop 13118 1727204086.24937: running the handler 13118 1727204086.25012: variable 'lsr_net_profile_exists' from source: set_fact 13118 1727204086.25015: Evaluated conditional (lsr_net_profile_exists): True 13118 1727204086.25022: handler run complete 13118 1727204086.25038: attempt loop complete, returning result 13118 1727204086.25041: _execute() done 13118 1727204086.25043: dumping result to json 13118 1727204086.25046: done dumping result, returning 13118 1727204086.25050: done running TaskExecutor() for managed-node2/TASK: Assert that the profile is present - 'bond0.0' [0affcd87-79f5-56a3-0a64-000000000267] 13118 1727204086.25053: sending task result for task 0affcd87-79f5-56a3-0a64-000000000267 13118 1727204086.25138: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000267 13118 1727204086.25145: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false } MSG: All assertions passed 13118 1727204086.25195: no more pending results, returning what we have 13118 1727204086.25199: results queue empty 13118 1727204086.25200: checking for any_errors_fatal 13118 1727204086.25206: done checking for any_errors_fatal 13118 1727204086.25207: checking for max_fail_percentage 13118 1727204086.25209: done checking for max_fail_percentage 13118 1727204086.25211: checking to see if all hosts have failed and the running result is not ok 13118 1727204086.25211: done checking to see if all hosts have failed 13118 1727204086.25212: getting the remaining hosts for this loop 13118 1727204086.25213: done getting the remaining hosts for this loop 13118 1727204086.25217: getting the next task for host managed-node2 13118 1727204086.25224: done getting next task for host managed-node2 13118 1727204086.25226: ^ task is: TASK: Assert that the ansible managed comment is present in '{{ profile }}' 13118 1727204086.25232: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204086.25236: getting variables 13118 1727204086.25238: in VariableManager get_vars() 13118 1727204086.25285: Calling all_inventory to load vars for managed-node2 13118 1727204086.25287: Calling groups_inventory to load vars for managed-node2 13118 1727204086.25289: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204086.25299: Calling all_plugins_play to load vars for managed-node2 13118 1727204086.25301: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204086.25304: Calling groups_plugins_play to load vars for managed-node2 13118 1727204086.26278: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204086.27206: done with get_vars() 13118 1727204086.27226: done getting variables 13118 1727204086.27275: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204086.27368: variable 'profile' from source: include params 13118 1727204086.27371: variable 'item' from source: include params 13118 1727204086.27412: variable 'item' from source: include params TASK [Assert that the ansible managed comment is present in 'bond0.0'] ********* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:10 Tuesday 24 September 2024 14:54:46 -0400 (0:00:00.040) 0:00:23.574 ***** 13118 1727204086.27445: entering _queue_task() for managed-node2/assert 13118 1727204086.27681: worker is 1 (out of 1 available) 13118 1727204086.27694: exiting _queue_task() for managed-node2/assert 13118 1727204086.27706: done queuing things up, now waiting for results queue to drain 13118 1727204086.27708: waiting for pending results... 13118 1727204086.27897: running TaskExecutor() for managed-node2/TASK: Assert that the ansible managed comment is present in 'bond0.0' 13118 1727204086.27970: in run() - task 0affcd87-79f5-56a3-0a64-000000000268 13118 1727204086.27981: variable 'ansible_search_path' from source: unknown 13118 1727204086.27985: variable 'ansible_search_path' from source: unknown 13118 1727204086.28014: calling self._execute() 13118 1727204086.28091: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.28095: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.28103: variable 'omit' from source: magic vars 13118 1727204086.28378: variable 'ansible_distribution_major_version' from source: facts 13118 1727204086.28388: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204086.28396: variable 'omit' from source: magic vars 13118 1727204086.28425: variable 'omit' from source: magic vars 13118 1727204086.28498: variable 'profile' from source: include params 13118 1727204086.28503: variable 'item' from source: include params 13118 1727204086.28551: variable 'item' from source: include params 13118 1727204086.28566: variable 'omit' from source: magic vars 13118 1727204086.28602: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204086.28635: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204086.28651: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204086.28666: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204086.28676: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204086.28700: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204086.28704: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.28706: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.28779: Set connection var ansible_timeout to 10 13118 1727204086.28788: Set connection var ansible_pipelining to False 13118 1727204086.28791: Set connection var ansible_connection to ssh 13118 1727204086.28796: Set connection var ansible_shell_executable to /bin/sh 13118 1727204086.28801: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204086.28804: Set connection var ansible_shell_type to sh 13118 1727204086.28824: variable 'ansible_shell_executable' from source: unknown 13118 1727204086.28827: variable 'ansible_connection' from source: unknown 13118 1727204086.28834: variable 'ansible_module_compression' from source: unknown 13118 1727204086.28836: variable 'ansible_shell_type' from source: unknown 13118 1727204086.28839: variable 'ansible_shell_executable' from source: unknown 13118 1727204086.28843: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.28845: variable 'ansible_pipelining' from source: unknown 13118 1727204086.28847: variable 'ansible_timeout' from source: unknown 13118 1727204086.28852: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.28957: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204086.28970: variable 'omit' from source: magic vars 13118 1727204086.28973: starting attempt loop 13118 1727204086.28976: running the handler 13118 1727204086.29056: variable 'lsr_net_profile_ansible_managed' from source: set_fact 13118 1727204086.29059: Evaluated conditional (lsr_net_profile_ansible_managed): True 13118 1727204086.29066: handler run complete 13118 1727204086.29078: attempt loop complete, returning result 13118 1727204086.29081: _execute() done 13118 1727204086.29084: dumping result to json 13118 1727204086.29087: done dumping result, returning 13118 1727204086.29092: done running TaskExecutor() for managed-node2/TASK: Assert that the ansible managed comment is present in 'bond0.0' [0affcd87-79f5-56a3-0a64-000000000268] 13118 1727204086.29098: sending task result for task 0affcd87-79f5-56a3-0a64-000000000268 13118 1727204086.29191: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000268 13118 1727204086.29193: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false } MSG: All assertions passed 13118 1727204086.29241: no more pending results, returning what we have 13118 1727204086.29245: results queue empty 13118 1727204086.29246: checking for any_errors_fatal 13118 1727204086.29252: done checking for any_errors_fatal 13118 1727204086.29252: checking for max_fail_percentage 13118 1727204086.29254: done checking for max_fail_percentage 13118 1727204086.29255: checking to see if all hosts have failed and the running result is not ok 13118 1727204086.29256: done checking to see if all hosts have failed 13118 1727204086.29257: getting the remaining hosts for this loop 13118 1727204086.29258: done getting the remaining hosts for this loop 13118 1727204086.29261: getting the next task for host managed-node2 13118 1727204086.29273: done getting next task for host managed-node2 13118 1727204086.29277: ^ task is: TASK: Assert that the fingerprint comment is present in {{ profile }} 13118 1727204086.29280: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204086.29284: getting variables 13118 1727204086.29285: in VariableManager get_vars() 13118 1727204086.29333: Calling all_inventory to load vars for managed-node2 13118 1727204086.29336: Calling groups_inventory to load vars for managed-node2 13118 1727204086.29338: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204086.29348: Calling all_plugins_play to load vars for managed-node2 13118 1727204086.29350: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204086.29353: Calling groups_plugins_play to load vars for managed-node2 13118 1727204086.30194: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204086.31126: done with get_vars() 13118 1727204086.31149: done getting variables 13118 1727204086.31197: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204086.31290: variable 'profile' from source: include params 13118 1727204086.31293: variable 'item' from source: include params 13118 1727204086.31337: variable 'item' from source: include params TASK [Assert that the fingerprint comment is present in bond0.0] *************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:15 Tuesday 24 September 2024 14:54:46 -0400 (0:00:00.039) 0:00:23.613 ***** 13118 1727204086.31368: entering _queue_task() for managed-node2/assert 13118 1727204086.31611: worker is 1 (out of 1 available) 13118 1727204086.31626: exiting _queue_task() for managed-node2/assert 13118 1727204086.31641: done queuing things up, now waiting for results queue to drain 13118 1727204086.31642: waiting for pending results... 13118 1727204086.31824: running TaskExecutor() for managed-node2/TASK: Assert that the fingerprint comment is present in bond0.0 13118 1727204086.31893: in run() - task 0affcd87-79f5-56a3-0a64-000000000269 13118 1727204086.31905: variable 'ansible_search_path' from source: unknown 13118 1727204086.31908: variable 'ansible_search_path' from source: unknown 13118 1727204086.31937: calling self._execute() 13118 1727204086.32015: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.32019: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.32027: variable 'omit' from source: magic vars 13118 1727204086.32357: variable 'ansible_distribution_major_version' from source: facts 13118 1727204086.32367: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204086.32374: variable 'omit' from source: magic vars 13118 1727204086.32400: variable 'omit' from source: magic vars 13118 1727204086.32475: variable 'profile' from source: include params 13118 1727204086.32479: variable 'item' from source: include params 13118 1727204086.32526: variable 'item' from source: include params 13118 1727204086.32541: variable 'omit' from source: magic vars 13118 1727204086.32577: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204086.32605: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204086.32622: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204086.32637: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204086.32648: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204086.32674: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204086.32677: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.32680: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.32749: Set connection var ansible_timeout to 10 13118 1727204086.32761: Set connection var ansible_pipelining to False 13118 1727204086.32764: Set connection var ansible_connection to ssh 13118 1727204086.32771: Set connection var ansible_shell_executable to /bin/sh 13118 1727204086.32774: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204086.32777: Set connection var ansible_shell_type to sh 13118 1727204086.32795: variable 'ansible_shell_executable' from source: unknown 13118 1727204086.32797: variable 'ansible_connection' from source: unknown 13118 1727204086.32800: variable 'ansible_module_compression' from source: unknown 13118 1727204086.32802: variable 'ansible_shell_type' from source: unknown 13118 1727204086.32804: variable 'ansible_shell_executable' from source: unknown 13118 1727204086.32806: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.32811: variable 'ansible_pipelining' from source: unknown 13118 1727204086.32814: variable 'ansible_timeout' from source: unknown 13118 1727204086.32818: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.32920: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204086.32929: variable 'omit' from source: magic vars 13118 1727204086.32934: starting attempt loop 13118 1727204086.32937: running the handler 13118 1727204086.33016: variable 'lsr_net_profile_fingerprint' from source: set_fact 13118 1727204086.33020: Evaluated conditional (lsr_net_profile_fingerprint): True 13118 1727204086.33026: handler run complete 13118 1727204086.33038: attempt loop complete, returning result 13118 1727204086.33041: _execute() done 13118 1727204086.33043: dumping result to json 13118 1727204086.33046: done dumping result, returning 13118 1727204086.33052: done running TaskExecutor() for managed-node2/TASK: Assert that the fingerprint comment is present in bond0.0 [0affcd87-79f5-56a3-0a64-000000000269] 13118 1727204086.33058: sending task result for task 0affcd87-79f5-56a3-0a64-000000000269 13118 1727204086.33143: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000269 13118 1727204086.33146: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false } MSG: All assertions passed 13118 1727204086.33199: no more pending results, returning what we have 13118 1727204086.33203: results queue empty 13118 1727204086.33204: checking for any_errors_fatal 13118 1727204086.33212: done checking for any_errors_fatal 13118 1727204086.33212: checking for max_fail_percentage 13118 1727204086.33214: done checking for max_fail_percentage 13118 1727204086.33215: checking to see if all hosts have failed and the running result is not ok 13118 1727204086.33216: done checking to see if all hosts have failed 13118 1727204086.33217: getting the remaining hosts for this loop 13118 1727204086.33218: done getting the remaining hosts for this loop 13118 1727204086.33222: getting the next task for host managed-node2 13118 1727204086.33234: done getting next task for host managed-node2 13118 1727204086.33236: ^ task is: TASK: Include the task 'get_profile_stat.yml' 13118 1727204086.33240: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204086.33244: getting variables 13118 1727204086.33246: in VariableManager get_vars() 13118 1727204086.33297: Calling all_inventory to load vars for managed-node2 13118 1727204086.33300: Calling groups_inventory to load vars for managed-node2 13118 1727204086.33302: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204086.33312: Calling all_plugins_play to load vars for managed-node2 13118 1727204086.33314: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204086.33317: Calling groups_plugins_play to load vars for managed-node2 13118 1727204086.35358: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204086.36535: done with get_vars() 13118 1727204086.36556: done getting variables TASK [Include the task 'get_profile_stat.yml'] ********************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:3 Tuesday 24 September 2024 14:54:46 -0400 (0:00:00.052) 0:00:23.666 ***** 13118 1727204086.36629: entering _queue_task() for managed-node2/include_tasks 13118 1727204086.36868: worker is 1 (out of 1 available) 13118 1727204086.36882: exiting _queue_task() for managed-node2/include_tasks 13118 1727204086.36896: done queuing things up, now waiting for results queue to drain 13118 1727204086.36897: waiting for pending results... 13118 1727204086.37095: running TaskExecutor() for managed-node2/TASK: Include the task 'get_profile_stat.yml' 13118 1727204086.37166: in run() - task 0affcd87-79f5-56a3-0a64-00000000026d 13118 1727204086.37178: variable 'ansible_search_path' from source: unknown 13118 1727204086.37183: variable 'ansible_search_path' from source: unknown 13118 1727204086.37214: calling self._execute() 13118 1727204086.37294: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.37298: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.37306: variable 'omit' from source: magic vars 13118 1727204086.38606: variable 'ansible_distribution_major_version' from source: facts 13118 1727204086.38772: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204086.38776: _execute() done 13118 1727204086.38778: dumping result to json 13118 1727204086.38780: done dumping result, returning 13118 1727204086.38782: done running TaskExecutor() for managed-node2/TASK: Include the task 'get_profile_stat.yml' [0affcd87-79f5-56a3-0a64-00000000026d] 13118 1727204086.38784: sending task result for task 0affcd87-79f5-56a3-0a64-00000000026d 13118 1727204086.38855: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000026d 13118 1727204086.38858: WORKER PROCESS EXITING 13118 1727204086.39004: no more pending results, returning what we have 13118 1727204086.39009: in VariableManager get_vars() 13118 1727204086.39054: Calling all_inventory to load vars for managed-node2 13118 1727204086.39057: Calling groups_inventory to load vars for managed-node2 13118 1727204086.39060: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204086.39072: Calling all_plugins_play to load vars for managed-node2 13118 1727204086.39075: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204086.39078: Calling groups_plugins_play to load vars for managed-node2 13118 1727204086.40479: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204086.42345: done with get_vars() 13118 1727204086.42367: variable 'ansible_search_path' from source: unknown 13118 1727204086.42369: variable 'ansible_search_path' from source: unknown 13118 1727204086.42409: we have included files to process 13118 1727204086.42410: generating all_blocks data 13118 1727204086.42412: done generating all_blocks data 13118 1727204086.42416: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13118 1727204086.42417: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13118 1727204086.42419: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13118 1727204086.43334: done processing included file 13118 1727204086.43336: iterating over new_blocks loaded from include file 13118 1727204086.43338: in VariableManager get_vars() 13118 1727204086.43361: done with get_vars() 13118 1727204086.43362: filtering new block on tags 13118 1727204086.43388: done filtering new block on tags 13118 1727204086.43391: in VariableManager get_vars() 13118 1727204086.43409: done with get_vars() 13118 1727204086.43410: filtering new block on tags 13118 1727204086.43433: done filtering new block on tags 13118 1727204086.43435: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml for managed-node2 13118 1727204086.43441: extending task lists for all hosts with included blocks 13118 1727204086.43621: done extending task lists 13118 1727204086.43622: done processing included files 13118 1727204086.43623: results queue empty 13118 1727204086.43624: checking for any_errors_fatal 13118 1727204086.43627: done checking for any_errors_fatal 13118 1727204086.43627: checking for max_fail_percentage 13118 1727204086.43629: done checking for max_fail_percentage 13118 1727204086.43631: checking to see if all hosts have failed and the running result is not ok 13118 1727204086.43632: done checking to see if all hosts have failed 13118 1727204086.43633: getting the remaining hosts for this loop 13118 1727204086.43634: done getting the remaining hosts for this loop 13118 1727204086.43637: getting the next task for host managed-node2 13118 1727204086.43641: done getting next task for host managed-node2 13118 1727204086.43643: ^ task is: TASK: Initialize NM profile exist and ansible_managed comment flag 13118 1727204086.43646: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204086.43648: getting variables 13118 1727204086.43649: in VariableManager get_vars() 13118 1727204086.43662: Calling all_inventory to load vars for managed-node2 13118 1727204086.43666: Calling groups_inventory to load vars for managed-node2 13118 1727204086.43669: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204086.43674: Calling all_plugins_play to load vars for managed-node2 13118 1727204086.43677: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204086.43679: Calling groups_plugins_play to load vars for managed-node2 13118 1727204086.44936: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204086.47093: done with get_vars() 13118 1727204086.47128: done getting variables 13118 1727204086.47185: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Initialize NM profile exist and ansible_managed comment flag] ************ task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:3 Tuesday 24 September 2024 14:54:46 -0400 (0:00:00.106) 0:00:23.772 ***** 13118 1727204086.47246: entering _queue_task() for managed-node2/set_fact 13118 1727204086.48175: worker is 1 (out of 1 available) 13118 1727204086.48251: exiting _queue_task() for managed-node2/set_fact 13118 1727204086.48468: done queuing things up, now waiting for results queue to drain 13118 1727204086.48470: waiting for pending results... 13118 1727204086.49941: running TaskExecutor() for managed-node2/TASK: Initialize NM profile exist and ansible_managed comment flag 13118 1727204086.50306: in run() - task 0affcd87-79f5-56a3-0a64-000000000440 13118 1727204086.50325: variable 'ansible_search_path' from source: unknown 13118 1727204086.50331: variable 'ansible_search_path' from source: unknown 13118 1727204086.50373: calling self._execute() 13118 1727204086.50474: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.50633: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.50649: variable 'omit' from source: magic vars 13118 1727204086.51351: variable 'ansible_distribution_major_version' from source: facts 13118 1727204086.51373: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204086.51390: variable 'omit' from source: magic vars 13118 1727204086.51538: variable 'omit' from source: magic vars 13118 1727204086.51583: variable 'omit' from source: magic vars 13118 1727204086.51644: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204086.51746: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204086.51838: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204086.51862: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204086.51940: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204086.51980: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204086.52044: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.52052: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.52275: Set connection var ansible_timeout to 10 13118 1727204086.52293: Set connection var ansible_pipelining to False 13118 1727204086.52301: Set connection var ansible_connection to ssh 13118 1727204086.52312: Set connection var ansible_shell_executable to /bin/sh 13118 1727204086.52323: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204086.52329: Set connection var ansible_shell_type to sh 13118 1727204086.52357: variable 'ansible_shell_executable' from source: unknown 13118 1727204086.52477: variable 'ansible_connection' from source: unknown 13118 1727204086.52486: variable 'ansible_module_compression' from source: unknown 13118 1727204086.52493: variable 'ansible_shell_type' from source: unknown 13118 1727204086.52500: variable 'ansible_shell_executable' from source: unknown 13118 1727204086.52507: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.52514: variable 'ansible_pipelining' from source: unknown 13118 1727204086.52521: variable 'ansible_timeout' from source: unknown 13118 1727204086.52531: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.52685: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204086.52821: variable 'omit' from source: magic vars 13118 1727204086.52924: starting attempt loop 13118 1727204086.52932: running the handler 13118 1727204086.52949: handler run complete 13118 1727204086.52966: attempt loop complete, returning result 13118 1727204086.52974: _execute() done 13118 1727204086.52981: dumping result to json 13118 1727204086.52988: done dumping result, returning 13118 1727204086.52998: done running TaskExecutor() for managed-node2/TASK: Initialize NM profile exist and ansible_managed comment flag [0affcd87-79f5-56a3-0a64-000000000440] 13118 1727204086.53008: sending task result for task 0affcd87-79f5-56a3-0a64-000000000440 ok: [managed-node2] => { "ansible_facts": { "lsr_net_profile_ansible_managed": false, "lsr_net_profile_exists": false, "lsr_net_profile_fingerprint": false }, "changed": false } 13118 1727204086.53171: no more pending results, returning what we have 13118 1727204086.53176: results queue empty 13118 1727204086.53177: checking for any_errors_fatal 13118 1727204086.53179: done checking for any_errors_fatal 13118 1727204086.53179: checking for max_fail_percentage 13118 1727204086.53181: done checking for max_fail_percentage 13118 1727204086.53182: checking to see if all hosts have failed and the running result is not ok 13118 1727204086.53182: done checking to see if all hosts have failed 13118 1727204086.53183: getting the remaining hosts for this loop 13118 1727204086.53184: done getting the remaining hosts for this loop 13118 1727204086.53189: getting the next task for host managed-node2 13118 1727204086.53196: done getting next task for host managed-node2 13118 1727204086.53198: ^ task is: TASK: Stat profile file 13118 1727204086.53202: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204086.53206: getting variables 13118 1727204086.53207: in VariableManager get_vars() 13118 1727204086.53254: Calling all_inventory to load vars for managed-node2 13118 1727204086.53257: Calling groups_inventory to load vars for managed-node2 13118 1727204086.53259: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204086.53275: Calling all_plugins_play to load vars for managed-node2 13118 1727204086.53278: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204086.53281: Calling groups_plugins_play to load vars for managed-node2 13118 1727204086.53800: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000440 13118 1727204086.53804: WORKER PROCESS EXITING 13118 1727204086.55854: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204086.60174: done with get_vars() 13118 1727204086.60208: done getting variables TASK [Stat profile file] ******************************************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:9 Tuesday 24 September 2024 14:54:46 -0400 (0:00:00.130) 0:00:23.903 ***** 13118 1727204086.60308: entering _queue_task() for managed-node2/stat 13118 1727204086.61048: worker is 1 (out of 1 available) 13118 1727204086.61062: exiting _queue_task() for managed-node2/stat 13118 1727204086.61077: done queuing things up, now waiting for results queue to drain 13118 1727204086.61079: waiting for pending results... 13118 1727204086.61754: running TaskExecutor() for managed-node2/TASK: Stat profile file 13118 1727204086.62501: in run() - task 0affcd87-79f5-56a3-0a64-000000000441 13118 1727204086.62523: variable 'ansible_search_path' from source: unknown 13118 1727204086.62531: variable 'ansible_search_path' from source: unknown 13118 1727204086.62576: calling self._execute() 13118 1727204086.62680: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.62693: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.62709: variable 'omit' from source: magic vars 13118 1727204086.63081: variable 'ansible_distribution_major_version' from source: facts 13118 1727204086.63786: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204086.63799: variable 'omit' from source: magic vars 13118 1727204086.63855: variable 'omit' from source: magic vars 13118 1727204086.63967: variable 'profile' from source: include params 13118 1727204086.63978: variable 'item' from source: include params 13118 1727204086.64048: variable 'item' from source: include params 13118 1727204086.64074: variable 'omit' from source: magic vars 13118 1727204086.64121: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204086.64162: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204086.64193: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204086.64216: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204086.64235: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204086.64274: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204086.64976: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.64984: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.65092: Set connection var ansible_timeout to 10 13118 1727204086.65110: Set connection var ansible_pipelining to False 13118 1727204086.65117: Set connection var ansible_connection to ssh 13118 1727204086.65127: Set connection var ansible_shell_executable to /bin/sh 13118 1727204086.65136: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204086.65143: Set connection var ansible_shell_type to sh 13118 1727204086.65173: variable 'ansible_shell_executable' from source: unknown 13118 1727204086.65181: variable 'ansible_connection' from source: unknown 13118 1727204086.65187: variable 'ansible_module_compression' from source: unknown 13118 1727204086.65193: variable 'ansible_shell_type' from source: unknown 13118 1727204086.65198: variable 'ansible_shell_executable' from source: unknown 13118 1727204086.65204: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204086.65213: variable 'ansible_pipelining' from source: unknown 13118 1727204086.65219: variable 'ansible_timeout' from source: unknown 13118 1727204086.65226: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204086.65429: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204086.65445: variable 'omit' from source: magic vars 13118 1727204086.65455: starting attempt loop 13118 1727204086.65461: running the handler 13118 1727204086.65482: _low_level_execute_command(): starting 13118 1727204086.65496: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204086.67284: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204086.67289: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204086.67401: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204086.67405: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204086.67408: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204086.67591: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204086.67595: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204086.67597: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204086.67655: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204086.69325: stdout chunk (state=3): >>>/root <<< 13118 1727204086.69426: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204086.69524: stderr chunk (state=3): >>><<< 13118 1727204086.69528: stdout chunk (state=3): >>><<< 13118 1727204086.69651: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204086.69655: _low_level_execute_command(): starting 13118 1727204086.69659: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204086.6955018-15201-117892665341224 `" && echo ansible-tmp-1727204086.6955018-15201-117892665341224="` echo /root/.ansible/tmp/ansible-tmp-1727204086.6955018-15201-117892665341224 `" ) && sleep 0' 13118 1727204086.72196: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204086.72201: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204086.72461: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204086.72475: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204086.72479: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204086.72527: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204086.72770: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204086.72779: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204086.72935: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204086.74831: stdout chunk (state=3): >>>ansible-tmp-1727204086.6955018-15201-117892665341224=/root/.ansible/tmp/ansible-tmp-1727204086.6955018-15201-117892665341224 <<< 13118 1727204086.74939: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204086.75030: stderr chunk (state=3): >>><<< 13118 1727204086.75034: stdout chunk (state=3): >>><<< 13118 1727204086.75171: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204086.6955018-15201-117892665341224=/root/.ansible/tmp/ansible-tmp-1727204086.6955018-15201-117892665341224 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204086.75175: variable 'ansible_module_compression' from source: unknown 13118 1727204086.75177: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13118 1727204086.75370: variable 'ansible_facts' from source: unknown 13118 1727204086.75373: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204086.6955018-15201-117892665341224/AnsiballZ_stat.py 13118 1727204086.75946: Sending initial data 13118 1727204086.75950: Sent initial data (153 bytes) 13118 1727204086.78052: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204086.78690: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204086.78711: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204086.78731: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204086.78781: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204086.78794: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204086.78810: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204086.78831: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204086.78843: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204086.78855: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204086.78872: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204086.78888: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204086.78905: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204086.78918: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204086.78929: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204086.78943: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204086.79023: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204086.79042: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204086.79057: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204086.79631: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204086.81471: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204086.81504: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204086.81544: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpno8pb723 /root/.ansible/tmp/ansible-tmp-1727204086.6955018-15201-117892665341224/AnsiballZ_stat.py <<< 13118 1727204086.81582: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204086.83270: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204086.83395: stderr chunk (state=3): >>><<< 13118 1727204086.83399: stdout chunk (state=3): >>><<< 13118 1727204086.83401: done transferring module to remote 13118 1727204086.83403: _low_level_execute_command(): starting 13118 1727204086.83406: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204086.6955018-15201-117892665341224/ /root/.ansible/tmp/ansible-tmp-1727204086.6955018-15201-117892665341224/AnsiballZ_stat.py && sleep 0' 13118 1727204086.84961: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204086.84967: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204086.85148: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204086.85152: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204086.85159: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204086.85161: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204086.85275: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204086.85279: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204086.85282: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204086.85389: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204086.87152: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204086.87236: stderr chunk (state=3): >>><<< 13118 1727204086.87240: stdout chunk (state=3): >>><<< 13118 1727204086.87334: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204086.87337: _low_level_execute_command(): starting 13118 1727204086.87340: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204086.6955018-15201-117892665341224/AnsiballZ_stat.py && sleep 0' 13118 1727204086.89129: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204086.89133: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204086.89173: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204086.89176: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204086.89179: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204086.89526: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204086.89529: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204086.89531: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204086.89623: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204087.02820: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/etc/sysconfig/network-scripts/ifcfg-bond0.1", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13118 1727204087.03807: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204087.03890: stderr chunk (state=3): >>><<< 13118 1727204087.03895: stdout chunk (state=3): >>><<< 13118 1727204087.03973: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/etc/sysconfig/network-scripts/ifcfg-bond0.1", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204087.03978: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/etc/sysconfig/network-scripts/ifcfg-bond0.1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204086.6955018-15201-117892665341224/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204087.03980: _low_level_execute_command(): starting 13118 1727204087.03982: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204086.6955018-15201-117892665341224/ > /dev/null 2>&1 && sleep 0' 13118 1727204087.04762: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204087.04784: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204087.04802: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204087.04821: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204087.04872: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204087.04885: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204087.04902: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204087.04919: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204087.04938: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204087.04949: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204087.04966: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204087.04983: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204087.05003: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204087.05015: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204087.05025: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204087.05040: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204087.05125: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204087.05144: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204087.05159: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204087.05230: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204087.07049: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204087.07102: stderr chunk (state=3): >>><<< 13118 1727204087.07105: stdout chunk (state=3): >>><<< 13118 1727204087.07121: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204087.07126: handler run complete 13118 1727204087.07145: attempt loop complete, returning result 13118 1727204087.07148: _execute() done 13118 1727204087.07152: dumping result to json 13118 1727204087.07154: done dumping result, returning 13118 1727204087.07162: done running TaskExecutor() for managed-node2/TASK: Stat profile file [0affcd87-79f5-56a3-0a64-000000000441] 13118 1727204087.07169: sending task result for task 0affcd87-79f5-56a3-0a64-000000000441 13118 1727204087.07263: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000441 13118 1727204087.07266: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "stat": { "exists": false } } 13118 1727204087.07322: no more pending results, returning what we have 13118 1727204087.07326: results queue empty 13118 1727204087.07327: checking for any_errors_fatal 13118 1727204087.07335: done checking for any_errors_fatal 13118 1727204087.07336: checking for max_fail_percentage 13118 1727204087.07338: done checking for max_fail_percentage 13118 1727204087.07339: checking to see if all hosts have failed and the running result is not ok 13118 1727204087.07340: done checking to see if all hosts have failed 13118 1727204087.07340: getting the remaining hosts for this loop 13118 1727204087.07342: done getting the remaining hosts for this loop 13118 1727204087.07345: getting the next task for host managed-node2 13118 1727204087.07353: done getting next task for host managed-node2 13118 1727204087.07355: ^ task is: TASK: Set NM profile exist flag based on the profile files 13118 1727204087.07359: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204087.07363: getting variables 13118 1727204087.07366: in VariableManager get_vars() 13118 1727204087.07409: Calling all_inventory to load vars for managed-node2 13118 1727204087.07412: Calling groups_inventory to load vars for managed-node2 13118 1727204087.07414: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204087.07426: Calling all_plugins_play to load vars for managed-node2 13118 1727204087.07428: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204087.07433: Calling groups_plugins_play to load vars for managed-node2 13118 1727204087.08485: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204087.09834: done with get_vars() 13118 1727204087.09857: done getting variables 13118 1727204087.09905: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Set NM profile exist flag based on the profile files] ******************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:17 Tuesday 24 September 2024 14:54:47 -0400 (0:00:00.496) 0:00:24.399 ***** 13118 1727204087.09933: entering _queue_task() for managed-node2/set_fact 13118 1727204087.10170: worker is 1 (out of 1 available) 13118 1727204087.10183: exiting _queue_task() for managed-node2/set_fact 13118 1727204087.10196: done queuing things up, now waiting for results queue to drain 13118 1727204087.10197: waiting for pending results... 13118 1727204087.10393: running TaskExecutor() for managed-node2/TASK: Set NM profile exist flag based on the profile files 13118 1727204087.10463: in run() - task 0affcd87-79f5-56a3-0a64-000000000442 13118 1727204087.10479: variable 'ansible_search_path' from source: unknown 13118 1727204087.10482: variable 'ansible_search_path' from source: unknown 13118 1727204087.10510: calling self._execute() 13118 1727204087.10593: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.10598: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.10607: variable 'omit' from source: magic vars 13118 1727204087.10889: variable 'ansible_distribution_major_version' from source: facts 13118 1727204087.10899: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204087.10986: variable 'profile_stat' from source: set_fact 13118 1727204087.10996: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204087.10999: when evaluation is False, skipping this task 13118 1727204087.11002: _execute() done 13118 1727204087.11005: dumping result to json 13118 1727204087.11009: done dumping result, returning 13118 1727204087.11018: done running TaskExecutor() for managed-node2/TASK: Set NM profile exist flag based on the profile files [0affcd87-79f5-56a3-0a64-000000000442] 13118 1727204087.11021: sending task result for task 0affcd87-79f5-56a3-0a64-000000000442 13118 1727204087.11114: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000442 13118 1727204087.11117: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204087.11168: no more pending results, returning what we have 13118 1727204087.11172: results queue empty 13118 1727204087.11173: checking for any_errors_fatal 13118 1727204087.11181: done checking for any_errors_fatal 13118 1727204087.11182: checking for max_fail_percentage 13118 1727204087.11183: done checking for max_fail_percentage 13118 1727204087.11184: checking to see if all hosts have failed and the running result is not ok 13118 1727204087.11185: done checking to see if all hosts have failed 13118 1727204087.11186: getting the remaining hosts for this loop 13118 1727204087.11187: done getting the remaining hosts for this loop 13118 1727204087.11191: getting the next task for host managed-node2 13118 1727204087.11198: done getting next task for host managed-node2 13118 1727204087.11200: ^ task is: TASK: Get NM profile info 13118 1727204087.11204: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204087.11210: getting variables 13118 1727204087.11212: in VariableManager get_vars() 13118 1727204087.11258: Calling all_inventory to load vars for managed-node2 13118 1727204087.11260: Calling groups_inventory to load vars for managed-node2 13118 1727204087.11262: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204087.11275: Calling all_plugins_play to load vars for managed-node2 13118 1727204087.11277: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204087.11280: Calling groups_plugins_play to load vars for managed-node2 13118 1727204087.12196: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204087.13113: done with get_vars() 13118 1727204087.13133: done getting variables 13118 1727204087.13179: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Get NM profile info] ***************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:25 Tuesday 24 September 2024 14:54:47 -0400 (0:00:00.032) 0:00:24.432 ***** 13118 1727204087.13205: entering _queue_task() for managed-node2/shell 13118 1727204087.13439: worker is 1 (out of 1 available) 13118 1727204087.13454: exiting _queue_task() for managed-node2/shell 13118 1727204087.13468: done queuing things up, now waiting for results queue to drain 13118 1727204087.13469: waiting for pending results... 13118 1727204087.13653: running TaskExecutor() for managed-node2/TASK: Get NM profile info 13118 1727204087.13725: in run() - task 0affcd87-79f5-56a3-0a64-000000000443 13118 1727204087.13736: variable 'ansible_search_path' from source: unknown 13118 1727204087.13740: variable 'ansible_search_path' from source: unknown 13118 1727204087.13770: calling self._execute() 13118 1727204087.13845: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.13849: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.13858: variable 'omit' from source: magic vars 13118 1727204087.14134: variable 'ansible_distribution_major_version' from source: facts 13118 1727204087.14144: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204087.14149: variable 'omit' from source: magic vars 13118 1727204087.14189: variable 'omit' from source: magic vars 13118 1727204087.14262: variable 'profile' from source: include params 13118 1727204087.14266: variable 'item' from source: include params 13118 1727204087.14312: variable 'item' from source: include params 13118 1727204087.14326: variable 'omit' from source: magic vars 13118 1727204087.14361: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204087.14392: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204087.14410: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204087.14424: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204087.14435: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204087.14458: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204087.14463: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.14465: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.14536: Set connection var ansible_timeout to 10 13118 1727204087.14543: Set connection var ansible_pipelining to False 13118 1727204087.14546: Set connection var ansible_connection to ssh 13118 1727204087.14551: Set connection var ansible_shell_executable to /bin/sh 13118 1727204087.14556: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204087.14559: Set connection var ansible_shell_type to sh 13118 1727204087.14578: variable 'ansible_shell_executable' from source: unknown 13118 1727204087.14580: variable 'ansible_connection' from source: unknown 13118 1727204087.14583: variable 'ansible_module_compression' from source: unknown 13118 1727204087.14585: variable 'ansible_shell_type' from source: unknown 13118 1727204087.14587: variable 'ansible_shell_executable' from source: unknown 13118 1727204087.14590: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.14594: variable 'ansible_pipelining' from source: unknown 13118 1727204087.14596: variable 'ansible_timeout' from source: unknown 13118 1727204087.14603: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.14700: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204087.14713: variable 'omit' from source: magic vars 13118 1727204087.14720: starting attempt loop 13118 1727204087.14723: running the handler 13118 1727204087.14726: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204087.14742: _low_level_execute_command(): starting 13118 1727204087.14749: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204087.15288: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204087.15298: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204087.15328: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204087.15347: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204087.15397: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204087.15410: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204087.15471: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204087.17109: stdout chunk (state=3): >>>/root <<< 13118 1727204087.17228: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204087.17287: stderr chunk (state=3): >>><<< 13118 1727204087.17290: stdout chunk (state=3): >>><<< 13118 1727204087.17314: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204087.17328: _low_level_execute_command(): starting 13118 1727204087.17336: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204087.1731246-15235-190791239254288 `" && echo ansible-tmp-1727204087.1731246-15235-190791239254288="` echo /root/.ansible/tmp/ansible-tmp-1727204087.1731246-15235-190791239254288 `" ) && sleep 0' 13118 1727204087.17805: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204087.17827: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204087.17844: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13118 1727204087.17861: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204087.17910: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204087.17915: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204087.17925: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204087.17984: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204087.19881: stdout chunk (state=3): >>>ansible-tmp-1727204087.1731246-15235-190791239254288=/root/.ansible/tmp/ansible-tmp-1727204087.1731246-15235-190791239254288 <<< 13118 1727204087.19997: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204087.20057: stderr chunk (state=3): >>><<< 13118 1727204087.20061: stdout chunk (state=3): >>><<< 13118 1727204087.20081: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204087.1731246-15235-190791239254288=/root/.ansible/tmp/ansible-tmp-1727204087.1731246-15235-190791239254288 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204087.20108: variable 'ansible_module_compression' from source: unknown 13118 1727204087.20152: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13118 1727204087.20186: variable 'ansible_facts' from source: unknown 13118 1727204087.20236: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204087.1731246-15235-190791239254288/AnsiballZ_command.py 13118 1727204087.20345: Sending initial data 13118 1727204087.20349: Sent initial data (156 bytes) 13118 1727204087.21213: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204087.21328: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204087.21334: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204087.21337: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204087.21339: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204087.21341: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204087.21343: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204087.21345: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204087.21470: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204087.21484: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204087.21497: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204087.21511: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204087.21528: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204087.21543: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204087.21554: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204087.21569: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204087.21649: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204087.21667: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204087.21683: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204087.21753: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204087.23481: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204087.23517: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204087.23553: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpx3hnum6v /root/.ansible/tmp/ansible-tmp-1727204087.1731246-15235-190791239254288/AnsiballZ_command.py <<< 13118 1727204087.23595: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204087.24576: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204087.24854: stderr chunk (state=3): >>><<< 13118 1727204087.24857: stdout chunk (state=3): >>><<< 13118 1727204087.24865: done transferring module to remote 13118 1727204087.24868: _low_level_execute_command(): starting 13118 1727204087.24870: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204087.1731246-15235-190791239254288/ /root/.ansible/tmp/ansible-tmp-1727204087.1731246-15235-190791239254288/AnsiballZ_command.py && sleep 0' 13118 1727204087.25474: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204087.25489: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204087.25505: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204087.25524: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204087.25574: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204087.25592: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204087.25606: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204087.25626: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204087.25642: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204087.25655: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204087.25672: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204087.25688: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204087.25704: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204087.25716: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204087.25728: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204087.25746: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204087.25823: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204087.25844: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204087.25858: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204087.26180: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204087.27703: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204087.27810: stderr chunk (state=3): >>><<< 13118 1727204087.27824: stdout chunk (state=3): >>><<< 13118 1727204087.27947: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204087.27952: _low_level_execute_command(): starting 13118 1727204087.27954: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204087.1731246-15235-190791239254288/AnsiballZ_command.py && sleep 0' 13118 1727204087.28608: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204087.28627: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204087.28646: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204087.28668: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204087.28744: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204087.28757: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204087.28785: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204087.28804: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204087.28818: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204087.28841: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204087.28866: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204087.28887: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204087.28904: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204087.28917: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204087.28929: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204087.28952: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204087.29032: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204087.29063: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204087.29084: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204087.29170: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204087.44778: stdout chunk (state=3): >>> {"changed": true, "stdout": "bond0.1 /etc/NetworkManager/system-connections/bond0.1.nmconnection ", "stderr": "", "rc": 0, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0.1 | grep /etc", "start": "2024-09-24 14:54:47.423775", "end": "2024-09-24 14:54:47.447098", "delta": "0:00:00.023323", "msg": "", "invocation": {"module_args": {"_raw_params": "nmcli -f NAME,FILENAME connection show |grep bond0.1 | grep /etc", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13118 1727204087.45945: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204087.46002: stderr chunk (state=3): >>><<< 13118 1727204087.46005: stdout chunk (state=3): >>><<< 13118 1727204087.46024: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "bond0.1 /etc/NetworkManager/system-connections/bond0.1.nmconnection ", "stderr": "", "rc": 0, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0.1 | grep /etc", "start": "2024-09-24 14:54:47.423775", "end": "2024-09-24 14:54:47.447098", "delta": "0:00:00.023323", "msg": "", "invocation": {"module_args": {"_raw_params": "nmcli -f NAME,FILENAME connection show |grep bond0.1 | grep /etc", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204087.46057: done with _execute_module (ansible.legacy.command, {'_raw_params': 'nmcli -f NAME,FILENAME connection show |grep bond0.1 | grep /etc', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204087.1731246-15235-190791239254288/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204087.46063: _low_level_execute_command(): starting 13118 1727204087.46074: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204087.1731246-15235-190791239254288/ > /dev/null 2>&1 && sleep 0' 13118 1727204087.46546: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204087.46550: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204087.46590: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204087.46594: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204087.46596: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204087.46654: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204087.46661: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204087.46667: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204087.46705: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204087.48480: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204087.48537: stderr chunk (state=3): >>><<< 13118 1727204087.48540: stdout chunk (state=3): >>><<< 13118 1727204087.48554: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204087.48562: handler run complete 13118 1727204087.48583: Evaluated conditional (False): False 13118 1727204087.48592: attempt loop complete, returning result 13118 1727204087.48594: _execute() done 13118 1727204087.48601: dumping result to json 13118 1727204087.48603: done dumping result, returning 13118 1727204087.48610: done running TaskExecutor() for managed-node2/TASK: Get NM profile info [0affcd87-79f5-56a3-0a64-000000000443] 13118 1727204087.48618: sending task result for task 0affcd87-79f5-56a3-0a64-000000000443 13118 1727204087.48717: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000443 13118 1727204087.48721: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0.1 | grep /etc", "delta": "0:00:00.023323", "end": "2024-09-24 14:54:47.447098", "rc": 0, "start": "2024-09-24 14:54:47.423775" } STDOUT: bond0.1 /etc/NetworkManager/system-connections/bond0.1.nmconnection 13118 1727204087.48790: no more pending results, returning what we have 13118 1727204087.48794: results queue empty 13118 1727204087.48795: checking for any_errors_fatal 13118 1727204087.48801: done checking for any_errors_fatal 13118 1727204087.48802: checking for max_fail_percentage 13118 1727204087.48804: done checking for max_fail_percentage 13118 1727204087.48805: checking to see if all hosts have failed and the running result is not ok 13118 1727204087.48805: done checking to see if all hosts have failed 13118 1727204087.48806: getting the remaining hosts for this loop 13118 1727204087.48807: done getting the remaining hosts for this loop 13118 1727204087.48812: getting the next task for host managed-node2 13118 1727204087.48820: done getting next task for host managed-node2 13118 1727204087.48822: ^ task is: TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output 13118 1727204087.48826: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204087.48832: getting variables 13118 1727204087.48833: in VariableManager get_vars() 13118 1727204087.48875: Calling all_inventory to load vars for managed-node2 13118 1727204087.48878: Calling groups_inventory to load vars for managed-node2 13118 1727204087.48880: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204087.48891: Calling all_plugins_play to load vars for managed-node2 13118 1727204087.48893: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204087.48895: Calling groups_plugins_play to load vars for managed-node2 13118 1727204087.49697: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204087.50712: done with get_vars() 13118 1727204087.50729: done getting variables 13118 1727204087.50775: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Set NM profile exist flag and ansible_managed flag true based on the nmcli output] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:35 Tuesday 24 September 2024 14:54:47 -0400 (0:00:00.375) 0:00:24.808 ***** 13118 1727204087.50802: entering _queue_task() for managed-node2/set_fact 13118 1727204087.51023: worker is 1 (out of 1 available) 13118 1727204087.51035: exiting _queue_task() for managed-node2/set_fact 13118 1727204087.51047: done queuing things up, now waiting for results queue to drain 13118 1727204087.51049: waiting for pending results... 13118 1727204087.51242: running TaskExecutor() for managed-node2/TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output 13118 1727204087.51308: in run() - task 0affcd87-79f5-56a3-0a64-000000000444 13118 1727204087.51319: variable 'ansible_search_path' from source: unknown 13118 1727204087.51322: variable 'ansible_search_path' from source: unknown 13118 1727204087.51352: calling self._execute() 13118 1727204087.51426: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.51430: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.51442: variable 'omit' from source: magic vars 13118 1727204087.51716: variable 'ansible_distribution_major_version' from source: facts 13118 1727204087.51726: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204087.51821: variable 'nm_profile_exists' from source: set_fact 13118 1727204087.51836: Evaluated conditional (nm_profile_exists.rc == 0): True 13118 1727204087.51842: variable 'omit' from source: magic vars 13118 1727204087.51876: variable 'omit' from source: magic vars 13118 1727204087.51898: variable 'omit' from source: magic vars 13118 1727204087.51936: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204087.51965: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204087.51981: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204087.51994: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204087.52003: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204087.52032: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204087.52035: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.52039: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.52108: Set connection var ansible_timeout to 10 13118 1727204087.52116: Set connection var ansible_pipelining to False 13118 1727204087.52120: Set connection var ansible_connection to ssh 13118 1727204087.52125: Set connection var ansible_shell_executable to /bin/sh 13118 1727204087.52129: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204087.52131: Set connection var ansible_shell_type to sh 13118 1727204087.52154: variable 'ansible_shell_executable' from source: unknown 13118 1727204087.52156: variable 'ansible_connection' from source: unknown 13118 1727204087.52159: variable 'ansible_module_compression' from source: unknown 13118 1727204087.52161: variable 'ansible_shell_type' from source: unknown 13118 1727204087.52163: variable 'ansible_shell_executable' from source: unknown 13118 1727204087.52167: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.52170: variable 'ansible_pipelining' from source: unknown 13118 1727204087.52172: variable 'ansible_timeout' from source: unknown 13118 1727204087.52176: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.52283: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204087.52292: variable 'omit' from source: magic vars 13118 1727204087.52297: starting attempt loop 13118 1727204087.52300: running the handler 13118 1727204087.52311: handler run complete 13118 1727204087.52318: attempt loop complete, returning result 13118 1727204087.52321: _execute() done 13118 1727204087.52324: dumping result to json 13118 1727204087.52326: done dumping result, returning 13118 1727204087.52336: done running TaskExecutor() for managed-node2/TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output [0affcd87-79f5-56a3-0a64-000000000444] 13118 1727204087.52343: sending task result for task 0affcd87-79f5-56a3-0a64-000000000444 13118 1727204087.52422: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000444 13118 1727204087.52425: WORKER PROCESS EXITING ok: [managed-node2] => { "ansible_facts": { "lsr_net_profile_ansible_managed": true, "lsr_net_profile_exists": true, "lsr_net_profile_fingerprint": true }, "changed": false } 13118 1727204087.52483: no more pending results, returning what we have 13118 1727204087.52487: results queue empty 13118 1727204087.52488: checking for any_errors_fatal 13118 1727204087.52497: done checking for any_errors_fatal 13118 1727204087.52498: checking for max_fail_percentage 13118 1727204087.52500: done checking for max_fail_percentage 13118 1727204087.52501: checking to see if all hosts have failed and the running result is not ok 13118 1727204087.52502: done checking to see if all hosts have failed 13118 1727204087.52503: getting the remaining hosts for this loop 13118 1727204087.52504: done getting the remaining hosts for this loop 13118 1727204087.52508: getting the next task for host managed-node2 13118 1727204087.52517: done getting next task for host managed-node2 13118 1727204087.52519: ^ task is: TASK: Get the ansible_managed comment in ifcfg-{{ profile }} 13118 1727204087.52523: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204087.52526: getting variables 13118 1727204087.52528: in VariableManager get_vars() 13118 1727204087.52578: Calling all_inventory to load vars for managed-node2 13118 1727204087.52580: Calling groups_inventory to load vars for managed-node2 13118 1727204087.52583: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204087.52591: Calling all_plugins_play to load vars for managed-node2 13118 1727204087.52594: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204087.52596: Calling groups_plugins_play to load vars for managed-node2 13118 1727204087.53378: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204087.54286: done with get_vars() 13118 1727204087.54304: done getting variables 13118 1727204087.54347: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204087.54437: variable 'profile' from source: include params 13118 1727204087.54440: variable 'item' from source: include params 13118 1727204087.54482: variable 'item' from source: include params TASK [Get the ansible_managed comment in ifcfg-bond0.1] ************************ task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:49 Tuesday 24 September 2024 14:54:47 -0400 (0:00:00.037) 0:00:24.845 ***** 13118 1727204087.54511: entering _queue_task() for managed-node2/command 13118 1727204087.54736: worker is 1 (out of 1 available) 13118 1727204087.54749: exiting _queue_task() for managed-node2/command 13118 1727204087.54762: done queuing things up, now waiting for results queue to drain 13118 1727204087.54763: waiting for pending results... 13118 1727204087.54954: running TaskExecutor() for managed-node2/TASK: Get the ansible_managed comment in ifcfg-bond0.1 13118 1727204087.55028: in run() - task 0affcd87-79f5-56a3-0a64-000000000446 13118 1727204087.55042: variable 'ansible_search_path' from source: unknown 13118 1727204087.55045: variable 'ansible_search_path' from source: unknown 13118 1727204087.55076: calling self._execute() 13118 1727204087.55149: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.55153: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.55161: variable 'omit' from source: magic vars 13118 1727204087.55422: variable 'ansible_distribution_major_version' from source: facts 13118 1727204087.55431: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204087.55517: variable 'profile_stat' from source: set_fact 13118 1727204087.55528: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204087.55533: when evaluation is False, skipping this task 13118 1727204087.55536: _execute() done 13118 1727204087.55539: dumping result to json 13118 1727204087.55545: done dumping result, returning 13118 1727204087.55548: done running TaskExecutor() for managed-node2/TASK: Get the ansible_managed comment in ifcfg-bond0.1 [0affcd87-79f5-56a3-0a64-000000000446] 13118 1727204087.55555: sending task result for task 0affcd87-79f5-56a3-0a64-000000000446 13118 1727204087.55637: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000446 13118 1727204087.55640: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204087.55696: no more pending results, returning what we have 13118 1727204087.55701: results queue empty 13118 1727204087.55702: checking for any_errors_fatal 13118 1727204087.55709: done checking for any_errors_fatal 13118 1727204087.55709: checking for max_fail_percentage 13118 1727204087.55711: done checking for max_fail_percentage 13118 1727204087.55712: checking to see if all hosts have failed and the running result is not ok 13118 1727204087.55713: done checking to see if all hosts have failed 13118 1727204087.55713: getting the remaining hosts for this loop 13118 1727204087.55715: done getting the remaining hosts for this loop 13118 1727204087.55718: getting the next task for host managed-node2 13118 1727204087.55726: done getting next task for host managed-node2 13118 1727204087.55727: ^ task is: TASK: Verify the ansible_managed comment in ifcfg-{{ profile }} 13118 1727204087.55731: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204087.55734: getting variables 13118 1727204087.55736: in VariableManager get_vars() 13118 1727204087.55779: Calling all_inventory to load vars for managed-node2 13118 1727204087.55782: Calling groups_inventory to load vars for managed-node2 13118 1727204087.55784: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204087.55793: Calling all_plugins_play to load vars for managed-node2 13118 1727204087.55795: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204087.55798: Calling groups_plugins_play to load vars for managed-node2 13118 1727204087.56707: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204087.57639: done with get_vars() 13118 1727204087.57655: done getting variables 13118 1727204087.57701: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204087.57787: variable 'profile' from source: include params 13118 1727204087.57790: variable 'item' from source: include params 13118 1727204087.57833: variable 'item' from source: include params TASK [Verify the ansible_managed comment in ifcfg-bond0.1] ********************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:56 Tuesday 24 September 2024 14:54:47 -0400 (0:00:00.033) 0:00:24.878 ***** 13118 1727204087.57857: entering _queue_task() for managed-node2/set_fact 13118 1727204087.58090: worker is 1 (out of 1 available) 13118 1727204087.58104: exiting _queue_task() for managed-node2/set_fact 13118 1727204087.58116: done queuing things up, now waiting for results queue to drain 13118 1727204087.58117: waiting for pending results... 13118 1727204087.58300: running TaskExecutor() for managed-node2/TASK: Verify the ansible_managed comment in ifcfg-bond0.1 13118 1727204087.58393: in run() - task 0affcd87-79f5-56a3-0a64-000000000447 13118 1727204087.58403: variable 'ansible_search_path' from source: unknown 13118 1727204087.58406: variable 'ansible_search_path' from source: unknown 13118 1727204087.58437: calling self._execute() 13118 1727204087.58514: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.58518: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.58525: variable 'omit' from source: magic vars 13118 1727204087.58857: variable 'ansible_distribution_major_version' from source: facts 13118 1727204087.58860: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204087.59025: variable 'profile_stat' from source: set_fact 13118 1727204087.59028: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204087.59030: when evaluation is False, skipping this task 13118 1727204087.59033: _execute() done 13118 1727204087.59035: dumping result to json 13118 1727204087.59037: done dumping result, returning 13118 1727204087.59040: done running TaskExecutor() for managed-node2/TASK: Verify the ansible_managed comment in ifcfg-bond0.1 [0affcd87-79f5-56a3-0a64-000000000447] 13118 1727204087.59042: sending task result for task 0affcd87-79f5-56a3-0a64-000000000447 13118 1727204087.59151: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000447 13118 1727204087.59153: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204087.59199: no more pending results, returning what we have 13118 1727204087.59204: results queue empty 13118 1727204087.59205: checking for any_errors_fatal 13118 1727204087.59213: done checking for any_errors_fatal 13118 1727204087.59213: checking for max_fail_percentage 13118 1727204087.59215: done checking for max_fail_percentage 13118 1727204087.59216: checking to see if all hosts have failed and the running result is not ok 13118 1727204087.59217: done checking to see if all hosts have failed 13118 1727204087.59217: getting the remaining hosts for this loop 13118 1727204087.59219: done getting the remaining hosts for this loop 13118 1727204087.59223: getting the next task for host managed-node2 13118 1727204087.59230: done getting next task for host managed-node2 13118 1727204087.59236: ^ task is: TASK: Get the fingerprint comment in ifcfg-{{ profile }} 13118 1727204087.59240: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204087.59243: getting variables 13118 1727204087.59245: in VariableManager get_vars() 13118 1727204087.59286: Calling all_inventory to load vars for managed-node2 13118 1727204087.59288: Calling groups_inventory to load vars for managed-node2 13118 1727204087.59290: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204087.59300: Calling all_plugins_play to load vars for managed-node2 13118 1727204087.59302: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204087.59305: Calling groups_plugins_play to load vars for managed-node2 13118 1727204087.60513: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204087.61466: done with get_vars() 13118 1727204087.61483: done getting variables 13118 1727204087.61527: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204087.61614: variable 'profile' from source: include params 13118 1727204087.61616: variable 'item' from source: include params 13118 1727204087.61658: variable 'item' from source: include params TASK [Get the fingerprint comment in ifcfg-bond0.1] **************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:62 Tuesday 24 September 2024 14:54:47 -0400 (0:00:00.038) 0:00:24.916 ***** 13118 1727204087.61683: entering _queue_task() for managed-node2/command 13118 1727204087.62613: worker is 1 (out of 1 available) 13118 1727204087.62622: exiting _queue_task() for managed-node2/command 13118 1727204087.62634: done queuing things up, now waiting for results queue to drain 13118 1727204087.62635: waiting for pending results... 13118 1727204087.62655: running TaskExecutor() for managed-node2/TASK: Get the fingerprint comment in ifcfg-bond0.1 13118 1727204087.62660: in run() - task 0affcd87-79f5-56a3-0a64-000000000448 13118 1727204087.62663: variable 'ansible_search_path' from source: unknown 13118 1727204087.62668: variable 'ansible_search_path' from source: unknown 13118 1727204087.62671: calling self._execute() 13118 1727204087.62673: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.62675: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.62678: variable 'omit' from source: magic vars 13118 1727204087.63096: variable 'ansible_distribution_major_version' from source: facts 13118 1727204087.63114: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204087.63278: variable 'profile_stat' from source: set_fact 13118 1727204087.63298: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204087.63305: when evaluation is False, skipping this task 13118 1727204087.63312: _execute() done 13118 1727204087.63319: dumping result to json 13118 1727204087.63326: done dumping result, returning 13118 1727204087.63338: done running TaskExecutor() for managed-node2/TASK: Get the fingerprint comment in ifcfg-bond0.1 [0affcd87-79f5-56a3-0a64-000000000448] 13118 1727204087.63348: sending task result for task 0affcd87-79f5-56a3-0a64-000000000448 13118 1727204087.63458: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000448 skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204087.63520: no more pending results, returning what we have 13118 1727204087.63526: results queue empty 13118 1727204087.63528: checking for any_errors_fatal 13118 1727204087.63537: done checking for any_errors_fatal 13118 1727204087.63538: checking for max_fail_percentage 13118 1727204087.63540: done checking for max_fail_percentage 13118 1727204087.63541: checking to see if all hosts have failed and the running result is not ok 13118 1727204087.63542: done checking to see if all hosts have failed 13118 1727204087.63543: getting the remaining hosts for this loop 13118 1727204087.63544: done getting the remaining hosts for this loop 13118 1727204087.63548: getting the next task for host managed-node2 13118 1727204087.63557: done getting next task for host managed-node2 13118 1727204087.63560: ^ task is: TASK: Verify the fingerprint comment in ifcfg-{{ profile }} 13118 1727204087.63567: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204087.63573: getting variables 13118 1727204087.63575: in VariableManager get_vars() 13118 1727204087.63624: Calling all_inventory to load vars for managed-node2 13118 1727204087.63628: Calling groups_inventory to load vars for managed-node2 13118 1727204087.63633: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204087.63647: Calling all_plugins_play to load vars for managed-node2 13118 1727204087.63650: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204087.63653: Calling groups_plugins_play to load vars for managed-node2 13118 1727204087.65049: WORKER PROCESS EXITING 13118 1727204087.72050: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204087.73743: done with get_vars() 13118 1727204087.73772: done getting variables 13118 1727204087.73820: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204087.73921: variable 'profile' from source: include params 13118 1727204087.73924: variable 'item' from source: include params 13118 1727204087.73987: variable 'item' from source: include params TASK [Verify the fingerprint comment in ifcfg-bond0.1] ************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:69 Tuesday 24 September 2024 14:54:47 -0400 (0:00:00.123) 0:00:25.040 ***** 13118 1727204087.74016: entering _queue_task() for managed-node2/set_fact 13118 1727204087.74346: worker is 1 (out of 1 available) 13118 1727204087.74359: exiting _queue_task() for managed-node2/set_fact 13118 1727204087.74372: done queuing things up, now waiting for results queue to drain 13118 1727204087.74374: waiting for pending results... 13118 1727204087.74666: running TaskExecutor() for managed-node2/TASK: Verify the fingerprint comment in ifcfg-bond0.1 13118 1727204087.74799: in run() - task 0affcd87-79f5-56a3-0a64-000000000449 13118 1727204087.74825: variable 'ansible_search_path' from source: unknown 13118 1727204087.74835: variable 'ansible_search_path' from source: unknown 13118 1727204087.74874: calling self._execute() 13118 1727204087.74977: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.74988: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.75001: variable 'omit' from source: magic vars 13118 1727204087.75371: variable 'ansible_distribution_major_version' from source: facts 13118 1727204087.75389: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204087.75515: variable 'profile_stat' from source: set_fact 13118 1727204087.75537: Evaluated conditional (profile_stat.stat.exists): False 13118 1727204087.75545: when evaluation is False, skipping this task 13118 1727204087.75551: _execute() done 13118 1727204087.75557: dumping result to json 13118 1727204087.75566: done dumping result, returning 13118 1727204087.75577: done running TaskExecutor() for managed-node2/TASK: Verify the fingerprint comment in ifcfg-bond0.1 [0affcd87-79f5-56a3-0a64-000000000449] 13118 1727204087.75587: sending task result for task 0affcd87-79f5-56a3-0a64-000000000449 skipping: [managed-node2] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13118 1727204087.75739: no more pending results, returning what we have 13118 1727204087.75744: results queue empty 13118 1727204087.75746: checking for any_errors_fatal 13118 1727204087.75753: done checking for any_errors_fatal 13118 1727204087.75754: checking for max_fail_percentage 13118 1727204087.75756: done checking for max_fail_percentage 13118 1727204087.75757: checking to see if all hosts have failed and the running result is not ok 13118 1727204087.75757: done checking to see if all hosts have failed 13118 1727204087.75758: getting the remaining hosts for this loop 13118 1727204087.75760: done getting the remaining hosts for this loop 13118 1727204087.75765: getting the next task for host managed-node2 13118 1727204087.75775: done getting next task for host managed-node2 13118 1727204087.75778: ^ task is: TASK: Assert that the profile is present - '{{ profile }}' 13118 1727204087.75781: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204087.75786: getting variables 13118 1727204087.75788: in VariableManager get_vars() 13118 1727204087.75829: Calling all_inventory to load vars for managed-node2 13118 1727204087.75835: Calling groups_inventory to load vars for managed-node2 13118 1727204087.75837: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204087.75851: Calling all_plugins_play to load vars for managed-node2 13118 1727204087.75854: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204087.75857: Calling groups_plugins_play to load vars for managed-node2 13118 1727204087.76884: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000449 13118 1727204087.76888: WORKER PROCESS EXITING 13118 1727204087.77558: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204087.80361: done with get_vars() 13118 1727204087.80392: done getting variables 13118 1727204087.80456: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204087.80582: variable 'profile' from source: include params 13118 1727204087.80585: variable 'item' from source: include params 13118 1727204087.80645: variable 'item' from source: include params TASK [Assert that the profile is present - 'bond0.1'] ************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:5 Tuesday 24 September 2024 14:54:47 -0400 (0:00:00.066) 0:00:25.106 ***** 13118 1727204087.80678: entering _queue_task() for managed-node2/assert 13118 1727204087.81005: worker is 1 (out of 1 available) 13118 1727204087.81019: exiting _queue_task() for managed-node2/assert 13118 1727204087.81033: done queuing things up, now waiting for results queue to drain 13118 1727204087.81034: waiting for pending results... 13118 1727204087.81322: running TaskExecutor() for managed-node2/TASK: Assert that the profile is present - 'bond0.1' 13118 1727204087.81441: in run() - task 0affcd87-79f5-56a3-0a64-00000000026e 13118 1727204087.81459: variable 'ansible_search_path' from source: unknown 13118 1727204087.81469: variable 'ansible_search_path' from source: unknown 13118 1727204087.81515: calling self._execute() 13118 1727204087.81625: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.81640: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.81653: variable 'omit' from source: magic vars 13118 1727204087.82034: variable 'ansible_distribution_major_version' from source: facts 13118 1727204087.82052: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204087.82063: variable 'omit' from source: magic vars 13118 1727204087.82113: variable 'omit' from source: magic vars 13118 1727204087.82227: variable 'profile' from source: include params 13118 1727204087.82244: variable 'item' from source: include params 13118 1727204087.82312: variable 'item' from source: include params 13118 1727204087.82342: variable 'omit' from source: magic vars 13118 1727204087.82399: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204087.82447: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204087.82481: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204087.82503: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204087.82519: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204087.82554: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204087.82563: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.82574: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.82679: Set connection var ansible_timeout to 10 13118 1727204087.82698: Set connection var ansible_pipelining to False 13118 1727204087.82704: Set connection var ansible_connection to ssh 13118 1727204087.82713: Set connection var ansible_shell_executable to /bin/sh 13118 1727204087.82723: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204087.82733: Set connection var ansible_shell_type to sh 13118 1727204087.82759: variable 'ansible_shell_executable' from source: unknown 13118 1727204087.82768: variable 'ansible_connection' from source: unknown 13118 1727204087.82775: variable 'ansible_module_compression' from source: unknown 13118 1727204087.82781: variable 'ansible_shell_type' from source: unknown 13118 1727204087.82788: variable 'ansible_shell_executable' from source: unknown 13118 1727204087.82797: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.82804: variable 'ansible_pipelining' from source: unknown 13118 1727204087.82811: variable 'ansible_timeout' from source: unknown 13118 1727204087.82818: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.82960: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204087.82979: variable 'omit' from source: magic vars 13118 1727204087.82989: starting attempt loop 13118 1727204087.82995: running the handler 13118 1727204087.83110: variable 'lsr_net_profile_exists' from source: set_fact 13118 1727204087.83124: Evaluated conditional (lsr_net_profile_exists): True 13118 1727204087.83137: handler run complete 13118 1727204087.83154: attempt loop complete, returning result 13118 1727204087.83160: _execute() done 13118 1727204087.83168: dumping result to json 13118 1727204087.83175: done dumping result, returning 13118 1727204087.83184: done running TaskExecutor() for managed-node2/TASK: Assert that the profile is present - 'bond0.1' [0affcd87-79f5-56a3-0a64-00000000026e] 13118 1727204087.83193: sending task result for task 0affcd87-79f5-56a3-0a64-00000000026e ok: [managed-node2] => { "changed": false } MSG: All assertions passed 13118 1727204087.83338: no more pending results, returning what we have 13118 1727204087.83342: results queue empty 13118 1727204087.83344: checking for any_errors_fatal 13118 1727204087.83351: done checking for any_errors_fatal 13118 1727204087.83352: checking for max_fail_percentage 13118 1727204087.83354: done checking for max_fail_percentage 13118 1727204087.83355: checking to see if all hosts have failed and the running result is not ok 13118 1727204087.83355: done checking to see if all hosts have failed 13118 1727204087.83356: getting the remaining hosts for this loop 13118 1727204087.83358: done getting the remaining hosts for this loop 13118 1727204087.83361: getting the next task for host managed-node2 13118 1727204087.83371: done getting next task for host managed-node2 13118 1727204087.83374: ^ task is: TASK: Assert that the ansible managed comment is present in '{{ profile }}' 13118 1727204087.83377: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204087.83381: getting variables 13118 1727204087.83383: in VariableManager get_vars() 13118 1727204087.83426: Calling all_inventory to load vars for managed-node2 13118 1727204087.83429: Calling groups_inventory to load vars for managed-node2 13118 1727204087.83434: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204087.83445: Calling all_plugins_play to load vars for managed-node2 13118 1727204087.83448: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204087.83451: Calling groups_plugins_play to load vars for managed-node2 13118 1727204087.84483: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000026e 13118 1727204087.84487: WORKER PROCESS EXITING 13118 1727204087.85340: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204087.86983: done with get_vars() 13118 1727204087.87009: done getting variables 13118 1727204087.87067: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204087.87179: variable 'profile' from source: include params 13118 1727204087.87182: variable 'item' from source: include params 13118 1727204087.87246: variable 'item' from source: include params TASK [Assert that the ansible managed comment is present in 'bond0.1'] ********* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:10 Tuesday 24 September 2024 14:54:47 -0400 (0:00:00.066) 0:00:25.172 ***** 13118 1727204087.87288: entering _queue_task() for managed-node2/assert 13118 1727204087.87635: worker is 1 (out of 1 available) 13118 1727204087.87649: exiting _queue_task() for managed-node2/assert 13118 1727204087.87661: done queuing things up, now waiting for results queue to drain 13118 1727204087.87663: waiting for pending results... 13118 1727204087.88753: running TaskExecutor() for managed-node2/TASK: Assert that the ansible managed comment is present in 'bond0.1' 13118 1727204087.88990: in run() - task 0affcd87-79f5-56a3-0a64-00000000026f 13118 1727204087.89014: variable 'ansible_search_path' from source: unknown 13118 1727204087.89022: variable 'ansible_search_path' from source: unknown 13118 1727204087.89182: calling self._execute() 13118 1727204087.89359: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.89374: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.89393: variable 'omit' from source: magic vars 13118 1727204087.90255: variable 'ansible_distribution_major_version' from source: facts 13118 1727204087.90399: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204087.90411: variable 'omit' from source: magic vars 13118 1727204087.90458: variable 'omit' from source: magic vars 13118 1727204087.90662: variable 'profile' from source: include params 13118 1727204087.90676: variable 'item' from source: include params 13118 1727204087.90751: variable 'item' from source: include params 13118 1727204087.90777: variable 'omit' from source: magic vars 13118 1727204087.90829: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204087.90875: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204087.90903: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204087.90932: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204087.90948: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204087.90984: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204087.90991: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.90998: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.91105: Set connection var ansible_timeout to 10 13118 1727204087.91119: Set connection var ansible_pipelining to False 13118 1727204087.91124: Set connection var ansible_connection to ssh 13118 1727204087.91133: Set connection var ansible_shell_executable to /bin/sh 13118 1727204087.91145: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204087.91150: Set connection var ansible_shell_type to sh 13118 1727204087.91175: variable 'ansible_shell_executable' from source: unknown 13118 1727204087.91181: variable 'ansible_connection' from source: unknown 13118 1727204087.91186: variable 'ansible_module_compression' from source: unknown 13118 1727204087.91191: variable 'ansible_shell_type' from source: unknown 13118 1727204087.91195: variable 'ansible_shell_executable' from source: unknown 13118 1727204087.91200: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.91206: variable 'ansible_pipelining' from source: unknown 13118 1727204087.91212: variable 'ansible_timeout' from source: unknown 13118 1727204087.91218: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.91364: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204087.91382: variable 'omit' from source: magic vars 13118 1727204087.91392: starting attempt loop 13118 1727204087.91399: running the handler 13118 1727204087.91517: variable 'lsr_net_profile_ansible_managed' from source: set_fact 13118 1727204087.91527: Evaluated conditional (lsr_net_profile_ansible_managed): True 13118 1727204087.91540: handler run complete 13118 1727204087.91557: attempt loop complete, returning result 13118 1727204087.91565: _execute() done 13118 1727204087.91573: dumping result to json 13118 1727204087.91583: done dumping result, returning 13118 1727204087.91593: done running TaskExecutor() for managed-node2/TASK: Assert that the ansible managed comment is present in 'bond0.1' [0affcd87-79f5-56a3-0a64-00000000026f] 13118 1727204087.91602: sending task result for task 0affcd87-79f5-56a3-0a64-00000000026f ok: [managed-node2] => { "changed": false } MSG: All assertions passed 13118 1727204087.91741: no more pending results, returning what we have 13118 1727204087.91745: results queue empty 13118 1727204087.91746: checking for any_errors_fatal 13118 1727204087.91754: done checking for any_errors_fatal 13118 1727204087.91755: checking for max_fail_percentage 13118 1727204087.91757: done checking for max_fail_percentage 13118 1727204087.91758: checking to see if all hosts have failed and the running result is not ok 13118 1727204087.91759: done checking to see if all hosts have failed 13118 1727204087.91760: getting the remaining hosts for this loop 13118 1727204087.91761: done getting the remaining hosts for this loop 13118 1727204087.91767: getting the next task for host managed-node2 13118 1727204087.91775: done getting next task for host managed-node2 13118 1727204087.91778: ^ task is: TASK: Assert that the fingerprint comment is present in {{ profile }} 13118 1727204087.91782: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204087.91786: getting variables 13118 1727204087.91788: in VariableManager get_vars() 13118 1727204087.91835: Calling all_inventory to load vars for managed-node2 13118 1727204087.91838: Calling groups_inventory to load vars for managed-node2 13118 1727204087.91840: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204087.91853: Calling all_plugins_play to load vars for managed-node2 13118 1727204087.91856: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204087.91859: Calling groups_plugins_play to load vars for managed-node2 13118 1727204087.92884: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000026f 13118 1727204087.92888: WORKER PROCESS EXITING 13118 1727204087.93599: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204087.95480: done with get_vars() 13118 1727204087.95504: done getting variables 13118 1727204087.95571: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204087.95689: variable 'profile' from source: include params 13118 1727204087.95693: variable 'item' from source: include params 13118 1727204087.95752: variable 'item' from source: include params TASK [Assert that the fingerprint comment is present in bond0.1] *************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:15 Tuesday 24 September 2024 14:54:47 -0400 (0:00:00.085) 0:00:25.257 ***** 13118 1727204087.95791: entering _queue_task() for managed-node2/assert 13118 1727204087.96113: worker is 1 (out of 1 available) 13118 1727204087.96124: exiting _queue_task() for managed-node2/assert 13118 1727204087.96138: done queuing things up, now waiting for results queue to drain 13118 1727204087.96140: waiting for pending results... 13118 1727204087.96437: running TaskExecutor() for managed-node2/TASK: Assert that the fingerprint comment is present in bond0.1 13118 1727204087.96567: in run() - task 0affcd87-79f5-56a3-0a64-000000000270 13118 1727204087.96591: variable 'ansible_search_path' from source: unknown 13118 1727204087.96599: variable 'ansible_search_path' from source: unknown 13118 1727204087.96639: calling self._execute() 13118 1727204087.96745: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.96756: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.96772: variable 'omit' from source: magic vars 13118 1727204087.97146: variable 'ansible_distribution_major_version' from source: facts 13118 1727204087.97165: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204087.97177: variable 'omit' from source: magic vars 13118 1727204087.97218: variable 'omit' from source: magic vars 13118 1727204087.97325: variable 'profile' from source: include params 13118 1727204087.97340: variable 'item' from source: include params 13118 1727204087.97396: variable 'item' from source: include params 13118 1727204087.97417: variable 'omit' from source: magic vars 13118 1727204087.97469: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204087.97511: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204087.97540: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204087.97570: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204087.97589: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204087.97625: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204087.97635: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.97641: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.97732: Set connection var ansible_timeout to 10 13118 1727204087.97748: Set connection var ansible_pipelining to False 13118 1727204087.97753: Set connection var ansible_connection to ssh 13118 1727204087.97761: Set connection var ansible_shell_executable to /bin/sh 13118 1727204087.97775: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204087.97780: Set connection var ansible_shell_type to sh 13118 1727204087.97803: variable 'ansible_shell_executable' from source: unknown 13118 1727204087.97809: variable 'ansible_connection' from source: unknown 13118 1727204087.97814: variable 'ansible_module_compression' from source: unknown 13118 1727204087.97819: variable 'ansible_shell_type' from source: unknown 13118 1727204087.97825: variable 'ansible_shell_executable' from source: unknown 13118 1727204087.97833: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204087.97843: variable 'ansible_pipelining' from source: unknown 13118 1727204087.97850: variable 'ansible_timeout' from source: unknown 13118 1727204087.97857: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204087.98011: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204087.98034: variable 'omit' from source: magic vars 13118 1727204087.98046: starting attempt loop 13118 1727204087.98053: running the handler 13118 1727204087.98174: variable 'lsr_net_profile_fingerprint' from source: set_fact 13118 1727204087.98185: Evaluated conditional (lsr_net_profile_fingerprint): True 13118 1727204087.98195: handler run complete 13118 1727204087.98219: attempt loop complete, returning result 13118 1727204087.98226: _execute() done 13118 1727204087.98236: dumping result to json 13118 1727204087.98244: done dumping result, returning 13118 1727204087.98254: done running TaskExecutor() for managed-node2/TASK: Assert that the fingerprint comment is present in bond0.1 [0affcd87-79f5-56a3-0a64-000000000270] 13118 1727204087.98266: sending task result for task 0affcd87-79f5-56a3-0a64-000000000270 ok: [managed-node2] => { "changed": false } MSG: All assertions passed 13118 1727204087.98424: no more pending results, returning what we have 13118 1727204087.98428: results queue empty 13118 1727204087.98432: checking for any_errors_fatal 13118 1727204087.98439: done checking for any_errors_fatal 13118 1727204087.98439: checking for max_fail_percentage 13118 1727204087.98441: done checking for max_fail_percentage 13118 1727204087.98442: checking to see if all hosts have failed and the running result is not ok 13118 1727204087.98443: done checking to see if all hosts have failed 13118 1727204087.98444: getting the remaining hosts for this loop 13118 1727204087.98446: done getting the remaining hosts for this loop 13118 1727204087.98449: getting the next task for host managed-node2 13118 1727204087.98459: done getting next task for host managed-node2 13118 1727204087.98462: ^ task is: TASK: ** TEST check polling interval 13118 1727204087.98466: ^ state is: HOST STATE: block=2, task=14, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204087.98470: getting variables 13118 1727204087.98472: in VariableManager get_vars() 13118 1727204087.98518: Calling all_inventory to load vars for managed-node2 13118 1727204087.98521: Calling groups_inventory to load vars for managed-node2 13118 1727204087.98524: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204087.98538: Calling all_plugins_play to load vars for managed-node2 13118 1727204087.98543: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204087.98547: Calling groups_plugins_play to load vars for managed-node2 13118 1727204088.00066: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000270 13118 1727204088.00071: WORKER PROCESS EXITING 13118 1727204088.01794: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204088.04261: done with get_vars() 13118 1727204088.04295: done getting variables 13118 1727204088.04360: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [** TEST check polling interval] ****************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml:75 Tuesday 24 September 2024 14:54:48 -0400 (0:00:00.085) 0:00:25.343 ***** 13118 1727204088.04393: entering _queue_task() for managed-node2/command 13118 1727204088.04737: worker is 1 (out of 1 available) 13118 1727204088.04749: exiting _queue_task() for managed-node2/command 13118 1727204088.04762: done queuing things up, now waiting for results queue to drain 13118 1727204088.04765: waiting for pending results... 13118 1727204088.05056: running TaskExecutor() for managed-node2/TASK: ** TEST check polling interval 13118 1727204088.05159: in run() - task 0affcd87-79f5-56a3-0a64-000000000071 13118 1727204088.05183: variable 'ansible_search_path' from source: unknown 13118 1727204088.05228: calling self._execute() 13118 1727204088.05347: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204088.05358: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204088.05374: variable 'omit' from source: magic vars 13118 1727204088.06005: variable 'ansible_distribution_major_version' from source: facts 13118 1727204088.06024: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204088.06039: variable 'omit' from source: magic vars 13118 1727204088.06063: variable 'omit' from source: magic vars 13118 1727204088.06174: variable 'controller_device' from source: play vars 13118 1727204088.06200: variable 'omit' from source: magic vars 13118 1727204088.06249: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204088.06294: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204088.06323: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204088.06347: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204088.06363: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204088.06401: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204088.06414: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204088.06421: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204088.06528: Set connection var ansible_timeout to 10 13118 1727204088.06548: Set connection var ansible_pipelining to False 13118 1727204088.06554: Set connection var ansible_connection to ssh 13118 1727204088.06562: Set connection var ansible_shell_executable to /bin/sh 13118 1727204088.06575: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204088.06583: Set connection var ansible_shell_type to sh 13118 1727204088.06612: variable 'ansible_shell_executable' from source: unknown 13118 1727204088.06622: variable 'ansible_connection' from source: unknown 13118 1727204088.06637: variable 'ansible_module_compression' from source: unknown 13118 1727204088.06645: variable 'ansible_shell_type' from source: unknown 13118 1727204088.06652: variable 'ansible_shell_executable' from source: unknown 13118 1727204088.06659: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204088.06670: variable 'ansible_pipelining' from source: unknown 13118 1727204088.06678: variable 'ansible_timeout' from source: unknown 13118 1727204088.06686: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204088.06835: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204088.06858: variable 'omit' from source: magic vars 13118 1727204088.06869: starting attempt loop 13118 1727204088.06876: running the handler 13118 1727204088.06893: _low_level_execute_command(): starting 13118 1727204088.06904: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204088.07723: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.07847: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.07862: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.07883: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.07942: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.07954: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.07970: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.07988: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.07999: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.08009: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.08021: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.08036: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.08054: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.08069: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.08081: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.08094: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.08178: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.08201: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.08216: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.08297: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.09990: stdout chunk (state=3): >>>/root <<< 13118 1727204088.10196: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204088.10200: stdout chunk (state=3): >>><<< 13118 1727204088.10203: stderr chunk (state=3): >>><<< 13118 1727204088.10336: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204088.10340: _low_level_execute_command(): starting 13118 1727204088.10343: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204088.1022727-15273-144672687372081 `" && echo ansible-tmp-1727204088.1022727-15273-144672687372081="` echo /root/.ansible/tmp/ansible-tmp-1727204088.1022727-15273-144672687372081 `" ) && sleep 0' 13118 1727204088.13112: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.13116: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.13156: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204088.13161: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found <<< 13118 1727204088.13165: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.13233: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.13237: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.13327: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.13361: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.15255: stdout chunk (state=3): >>>ansible-tmp-1727204088.1022727-15273-144672687372081=/root/.ansible/tmp/ansible-tmp-1727204088.1022727-15273-144672687372081 <<< 13118 1727204088.15356: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204088.15440: stderr chunk (state=3): >>><<< 13118 1727204088.15444: stdout chunk (state=3): >>><<< 13118 1727204088.15772: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204088.1022727-15273-144672687372081=/root/.ansible/tmp/ansible-tmp-1727204088.1022727-15273-144672687372081 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204088.15776: variable 'ansible_module_compression' from source: unknown 13118 1727204088.15778: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13118 1727204088.15780: variable 'ansible_facts' from source: unknown 13118 1727204088.15782: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204088.1022727-15273-144672687372081/AnsiballZ_command.py 13118 1727204088.16326: Sending initial data 13118 1727204088.16333: Sent initial data (156 bytes) 13118 1727204088.17341: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.17353: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.17363: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.17378: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.17416: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.17424: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.17435: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.17448: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.17460: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.17469: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.17477: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.17486: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.17498: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.17506: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.17512: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.17521: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.17619: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.17638: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.17666: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.17713: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.19496: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204088.19555: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204088.19593: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmp9uyydrsn /root/.ansible/tmp/ansible-tmp-1727204088.1022727-15273-144672687372081/AnsiballZ_command.py <<< 13118 1727204088.19607: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204088.21070: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204088.21074: stderr chunk (state=3): >>><<< 13118 1727204088.21077: stdout chunk (state=3): >>><<< 13118 1727204088.21079: done transferring module to remote 13118 1727204088.21081: _low_level_execute_command(): starting 13118 1727204088.21083: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204088.1022727-15273-144672687372081/ /root/.ansible/tmp/ansible-tmp-1727204088.1022727-15273-144672687372081/AnsiballZ_command.py && sleep 0' 13118 1727204088.22238: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.22253: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.22270: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.22297: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.22342: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.22355: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.22372: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.22389: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.22402: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.22416: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.22428: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.22444: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.22458: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.22472: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.22483: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.22495: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.22577: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.22598: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.22612: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.22686: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.24494: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204088.24561: stderr chunk (state=3): >>><<< 13118 1727204088.24568: stdout chunk (state=3): >>><<< 13118 1727204088.24676: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204088.24679: _low_level_execute_command(): starting 13118 1727204088.24682: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204088.1022727-15273-144672687372081/AnsiballZ_command.py && sleep 0' 13118 1727204088.25421: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.25434: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.25444: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.25454: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.25496: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.25502: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.25513: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.25526: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.25535: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.25540: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.25549: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.25559: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.25571: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.25579: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.25585: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.25595: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.25666: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.25684: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.25692: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.25793: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.39348: stdout chunk (state=3): >>> {"changed": true, "stdout": "MII Polling Interval (ms): 110", "stderr": "", "rc": 0, "cmd": ["grep", "Polling Interval", "/proc/net/bonding/nm-bond"], "start": "2024-09-24 14:54:48.389217", "end": "2024-09-24 14:54:48.392656", "delta": "0:00:00.003439", "msg": "", "invocation": {"module_args": {"_raw_params": "grep 'Polling Interval' /proc/net/bonding/nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13118 1727204088.40623: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204088.40627: stdout chunk (state=3): >>><<< 13118 1727204088.40630: stderr chunk (state=3): >>><<< 13118 1727204088.40770: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "MII Polling Interval (ms): 110", "stderr": "", "rc": 0, "cmd": ["grep", "Polling Interval", "/proc/net/bonding/nm-bond"], "start": "2024-09-24 14:54:48.389217", "end": "2024-09-24 14:54:48.392656", "delta": "0:00:00.003439", "msg": "", "invocation": {"module_args": {"_raw_params": "grep 'Polling Interval' /proc/net/bonding/nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204088.40780: done with _execute_module (ansible.legacy.command, {'_raw_params': "grep 'Polling Interval' /proc/net/bonding/nm-bond", '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204088.1022727-15273-144672687372081/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204088.40783: _low_level_execute_command(): starting 13118 1727204088.40786: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204088.1022727-15273-144672687372081/ > /dev/null 2>&1 && sleep 0' 13118 1727204088.41365: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.41382: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.41398: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.41417: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.41462: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.41479: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.41492: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.41509: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.41522: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.41534: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.41546: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.41558: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.41575: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.41586: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.41596: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.41610: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.41692: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.41708: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.41723: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.41800: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.43690: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204088.43694: stdout chunk (state=3): >>><<< 13118 1727204088.43703: stderr chunk (state=3): >>><<< 13118 1727204088.43721: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204088.43727: handler run complete 13118 1727204088.43757: Evaluated conditional (False): False 13118 1727204088.43916: variable 'result' from source: unknown 13118 1727204088.43933: Evaluated conditional ('110' in result.stdout): True 13118 1727204088.43951: attempt loop complete, returning result 13118 1727204088.43954: _execute() done 13118 1727204088.43956: dumping result to json 13118 1727204088.43963: done dumping result, returning 13118 1727204088.43973: done running TaskExecutor() for managed-node2/TASK: ** TEST check polling interval [0affcd87-79f5-56a3-0a64-000000000071] 13118 1727204088.43979: sending task result for task 0affcd87-79f5-56a3-0a64-000000000071 13118 1727204088.44094: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000071 13118 1727204088.44097: WORKER PROCESS EXITING ok: [managed-node2] => { "attempts": 1, "changed": false, "cmd": [ "grep", "Polling Interval", "/proc/net/bonding/nm-bond" ], "delta": "0:00:00.003439", "end": "2024-09-24 14:54:48.392656", "rc": 0, "start": "2024-09-24 14:54:48.389217" } STDOUT: MII Polling Interval (ms): 110 13118 1727204088.44166: no more pending results, returning what we have 13118 1727204088.44170: results queue empty 13118 1727204088.44171: checking for any_errors_fatal 13118 1727204088.44178: done checking for any_errors_fatal 13118 1727204088.44178: checking for max_fail_percentage 13118 1727204088.44181: done checking for max_fail_percentage 13118 1727204088.44182: checking to see if all hosts have failed and the running result is not ok 13118 1727204088.44183: done checking to see if all hosts have failed 13118 1727204088.44183: getting the remaining hosts for this loop 13118 1727204088.44185: done getting the remaining hosts for this loop 13118 1727204088.44189: getting the next task for host managed-node2 13118 1727204088.44195: done getting next task for host managed-node2 13118 1727204088.44197: ^ task is: TASK: ** TEST check IPv4 13118 1727204088.44199: ^ state is: HOST STATE: block=2, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204088.44203: getting variables 13118 1727204088.44204: in VariableManager get_vars() 13118 1727204088.44245: Calling all_inventory to load vars for managed-node2 13118 1727204088.44248: Calling groups_inventory to load vars for managed-node2 13118 1727204088.44250: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204088.44260: Calling all_plugins_play to load vars for managed-node2 13118 1727204088.44262: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204088.44267: Calling groups_plugins_play to load vars for managed-node2 13118 1727204088.45847: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204088.47598: done with get_vars() 13118 1727204088.47631: done getting variables 13118 1727204088.47710: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [** TEST check IPv4] ****************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml:80 Tuesday 24 September 2024 14:54:48 -0400 (0:00:00.433) 0:00:25.777 ***** 13118 1727204088.47750: entering _queue_task() for managed-node2/command 13118 1727204088.48199: worker is 1 (out of 1 available) 13118 1727204088.48233: exiting _queue_task() for managed-node2/command 13118 1727204088.48253: done queuing things up, now waiting for results queue to drain 13118 1727204088.48255: waiting for pending results... 13118 1727204088.48699: running TaskExecutor() for managed-node2/TASK: ** TEST check IPv4 13118 1727204088.48804: in run() - task 0affcd87-79f5-56a3-0a64-000000000072 13118 1727204088.48830: variable 'ansible_search_path' from source: unknown 13118 1727204088.48876: calling self._execute() 13118 1727204088.48982: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204088.48998: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204088.49011: variable 'omit' from source: magic vars 13118 1727204088.49383: variable 'ansible_distribution_major_version' from source: facts 13118 1727204088.49399: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204088.49409: variable 'omit' from source: magic vars 13118 1727204088.49438: variable 'omit' from source: magic vars 13118 1727204088.49558: variable 'controller_device' from source: play vars 13118 1727204088.49585: variable 'omit' from source: magic vars 13118 1727204088.49633: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204088.49679: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204088.49708: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204088.49729: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204088.49746: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204088.49787: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204088.49799: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204088.49806: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204088.49915: Set connection var ansible_timeout to 10 13118 1727204088.49931: Set connection var ansible_pipelining to False 13118 1727204088.49937: Set connection var ansible_connection to ssh 13118 1727204088.49946: Set connection var ansible_shell_executable to /bin/sh 13118 1727204088.49954: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204088.49961: Set connection var ansible_shell_type to sh 13118 1727204088.49993: variable 'ansible_shell_executable' from source: unknown 13118 1727204088.49999: variable 'ansible_connection' from source: unknown 13118 1727204088.50006: variable 'ansible_module_compression' from source: unknown 13118 1727204088.50016: variable 'ansible_shell_type' from source: unknown 13118 1727204088.50022: variable 'ansible_shell_executable' from source: unknown 13118 1727204088.50028: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204088.50035: variable 'ansible_pipelining' from source: unknown 13118 1727204088.50041: variable 'ansible_timeout' from source: unknown 13118 1727204088.50048: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204088.50197: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204088.50215: variable 'omit' from source: magic vars 13118 1727204088.50224: starting attempt loop 13118 1727204088.50238: running the handler 13118 1727204088.50259: _low_level_execute_command(): starting 13118 1727204088.50274: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204088.51079: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.51124: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.51140: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.51158: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.51209: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.51224: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.51238: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.51256: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.51268: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.51279: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.51295: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.51308: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.51323: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.51338: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.51349: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.51362: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.51446: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.51471: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.51490: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.51570: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.53177: stdout chunk (state=3): >>>/root <<< 13118 1727204088.53370: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204088.53374: stdout chunk (state=3): >>><<< 13118 1727204088.53409: stderr chunk (state=3): >>><<< 13118 1727204088.53536: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204088.53540: _low_level_execute_command(): starting 13118 1727204088.53543: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204088.5343306-15298-13486172099263 `" && echo ansible-tmp-1727204088.5343306-15298-13486172099263="` echo /root/.ansible/tmp/ansible-tmp-1727204088.5343306-15298-13486172099263 `" ) && sleep 0' 13118 1727204088.54278: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.54294: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.54317: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.54337: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.54383: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.54395: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.54416: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.54436: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.54446: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.54455: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.54466: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.54476: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.54489: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.54498: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.54506: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.54516: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.54594: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.54615: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.54634: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.54714: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.56569: stdout chunk (state=3): >>>ansible-tmp-1727204088.5343306-15298-13486172099263=/root/.ansible/tmp/ansible-tmp-1727204088.5343306-15298-13486172099263 <<< 13118 1727204088.56768: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204088.56772: stdout chunk (state=3): >>><<< 13118 1727204088.56774: stderr chunk (state=3): >>><<< 13118 1727204088.57074: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204088.5343306-15298-13486172099263=/root/.ansible/tmp/ansible-tmp-1727204088.5343306-15298-13486172099263 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204088.57078: variable 'ansible_module_compression' from source: unknown 13118 1727204088.57080: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13118 1727204088.57082: variable 'ansible_facts' from source: unknown 13118 1727204088.57084: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204088.5343306-15298-13486172099263/AnsiballZ_command.py 13118 1727204088.57204: Sending initial data 13118 1727204088.57207: Sent initial data (155 bytes) 13118 1727204088.58142: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.58157: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.58182: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.58202: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.58244: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.58255: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.58275: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.58292: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.58302: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.58312: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.58322: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.58371: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.58393: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.58404: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.58558: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.58717: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.58794: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.58819: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.58835: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.58906: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.60658: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204088.60728: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204088.60732: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpeg638opt /root/.ansible/tmp/ansible-tmp-1727204088.5343306-15298-13486172099263/AnsiballZ_command.py <<< 13118 1727204088.61033: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204088.61854: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204088.61986: stderr chunk (state=3): >>><<< 13118 1727204088.61989: stdout chunk (state=3): >>><<< 13118 1727204088.62085: done transferring module to remote 13118 1727204088.62092: _low_level_execute_command(): starting 13118 1727204088.62095: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204088.5343306-15298-13486172099263/ /root/.ansible/tmp/ansible-tmp-1727204088.5343306-15298-13486172099263/AnsiballZ_command.py && sleep 0' 13118 1727204088.63127: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.63788: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.63792: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.63802: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.63840: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.63852: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.63855: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.63881: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.63884: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.63887: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.63891: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.63897: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.63919: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.63922: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.63924: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.63930: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.64007: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.64088: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.64101: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.64265: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.66072: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204088.66222: stderr chunk (state=3): >>><<< 13118 1727204088.66226: stdout chunk (state=3): >>><<< 13118 1727204088.67143: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204088.67147: _low_level_execute_command(): starting 13118 1727204088.67150: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204088.5343306-15298-13486172099263/AnsiballZ_command.py && sleep 0' 13118 1727204088.68073: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.68077: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.68088: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.68100: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.68140: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.68153: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.68172: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.68186: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.68194: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.68201: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.68209: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.68218: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.68233: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.68237: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.68245: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.68260: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.68336: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.68355: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.68373: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.68451: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.82053: stdout chunk (state=3): >>> {"changed": true, "stdout": "13: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000\n inet 192.0.2.14/24 brd 192.0.2.255 scope global dynamic noprefixroute nm-bond\n valid_lft 233sec preferred_lft 233sec", "stderr": "", "rc": 0, "cmd": ["ip", "-4", "a", "s", "nm-bond"], "start": "2024-09-24 14:54:48.815814", "end": "2024-09-24 14:54:48.819723", "delta": "0:00:00.003909", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -4 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13118 1727204088.83229: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204088.83289: stderr chunk (state=3): >>><<< 13118 1727204088.83293: stdout chunk (state=3): >>><<< 13118 1727204088.83308: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "13: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000\n inet 192.0.2.14/24 brd 192.0.2.255 scope global dynamic noprefixroute nm-bond\n valid_lft 233sec preferred_lft 233sec", "stderr": "", "rc": 0, "cmd": ["ip", "-4", "a", "s", "nm-bond"], "start": "2024-09-24 14:54:48.815814", "end": "2024-09-24 14:54:48.819723", "delta": "0:00:00.003909", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -4 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204088.83340: done with _execute_module (ansible.legacy.command, {'_raw_params': 'ip -4 a s nm-bond', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204088.5343306-15298-13486172099263/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204088.83349: _low_level_execute_command(): starting 13118 1727204088.83354: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204088.5343306-15298-13486172099263/ > /dev/null 2>&1 && sleep 0' 13118 1727204088.83815: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.83821: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.83868: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.83872: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.83874: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.83937: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.83944: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.83945: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.83984: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.85761: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204088.85811: stderr chunk (state=3): >>><<< 13118 1727204088.85816: stdout chunk (state=3): >>><<< 13118 1727204088.85834: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204088.85837: handler run complete 13118 1727204088.85857: Evaluated conditional (False): False 13118 1727204088.85972: variable 'result' from source: set_fact 13118 1727204088.85985: Evaluated conditional ('192.0.2' in result.stdout): True 13118 1727204088.85995: attempt loop complete, returning result 13118 1727204088.85998: _execute() done 13118 1727204088.86000: dumping result to json 13118 1727204088.86006: done dumping result, returning 13118 1727204088.86013: done running TaskExecutor() for managed-node2/TASK: ** TEST check IPv4 [0affcd87-79f5-56a3-0a64-000000000072] 13118 1727204088.86019: sending task result for task 0affcd87-79f5-56a3-0a64-000000000072 13118 1727204088.86116: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000072 13118 1727204088.86119: WORKER PROCESS EXITING ok: [managed-node2] => { "attempts": 1, "changed": false, "cmd": [ "ip", "-4", "a", "s", "nm-bond" ], "delta": "0:00:00.003909", "end": "2024-09-24 14:54:48.819723", "rc": 0, "start": "2024-09-24 14:54:48.815814" } STDOUT: 13: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000 inet 192.0.2.14/24 brd 192.0.2.255 scope global dynamic noprefixroute nm-bond valid_lft 233sec preferred_lft 233sec 13118 1727204088.86194: no more pending results, returning what we have 13118 1727204088.86198: results queue empty 13118 1727204088.86199: checking for any_errors_fatal 13118 1727204088.86209: done checking for any_errors_fatal 13118 1727204088.86210: checking for max_fail_percentage 13118 1727204088.86212: done checking for max_fail_percentage 13118 1727204088.86213: checking to see if all hosts have failed and the running result is not ok 13118 1727204088.86214: done checking to see if all hosts have failed 13118 1727204088.86214: getting the remaining hosts for this loop 13118 1727204088.86216: done getting the remaining hosts for this loop 13118 1727204088.86219: getting the next task for host managed-node2 13118 1727204088.86225: done getting next task for host managed-node2 13118 1727204088.86228: ^ task is: TASK: ** TEST check IPv6 13118 1727204088.86229: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204088.86236: getting variables 13118 1727204088.86237: in VariableManager get_vars() 13118 1727204088.86277: Calling all_inventory to load vars for managed-node2 13118 1727204088.86280: Calling groups_inventory to load vars for managed-node2 13118 1727204088.86282: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204088.86292: Calling all_plugins_play to load vars for managed-node2 13118 1727204088.86294: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204088.86297: Calling groups_plugins_play to load vars for managed-node2 13118 1727204088.87754: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204088.88681: done with get_vars() 13118 1727204088.88697: done getting variables 13118 1727204088.88743: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [** TEST check IPv6] ****************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml:87 Tuesday 24 September 2024 14:54:48 -0400 (0:00:00.410) 0:00:26.187 ***** 13118 1727204088.88766: entering _queue_task() for managed-node2/command 13118 1727204088.88981: worker is 1 (out of 1 available) 13118 1727204088.88994: exiting _queue_task() for managed-node2/command 13118 1727204088.89007: done queuing things up, now waiting for results queue to drain 13118 1727204088.89008: waiting for pending results... 13118 1727204088.89201: running TaskExecutor() for managed-node2/TASK: ** TEST check IPv6 13118 1727204088.89293: in run() - task 0affcd87-79f5-56a3-0a64-000000000073 13118 1727204088.89312: variable 'ansible_search_path' from source: unknown 13118 1727204088.89353: calling self._execute() 13118 1727204088.89465: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204088.89476: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204088.89489: variable 'omit' from source: magic vars 13118 1727204088.90170: variable 'ansible_distribution_major_version' from source: facts 13118 1727204088.90173: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204088.90176: variable 'omit' from source: magic vars 13118 1727204088.90180: variable 'omit' from source: magic vars 13118 1727204088.90182: variable 'controller_device' from source: play vars 13118 1727204088.90184: variable 'omit' from source: magic vars 13118 1727204088.90187: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204088.90189: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204088.90191: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204088.90193: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204088.90196: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204088.90198: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204088.90200: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204088.90202: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204088.90283: Set connection var ansible_timeout to 10 13118 1727204088.90294: Set connection var ansible_pipelining to False 13118 1727204088.90296: Set connection var ansible_connection to ssh 13118 1727204088.90302: Set connection var ansible_shell_executable to /bin/sh 13118 1727204088.90307: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204088.90310: Set connection var ansible_shell_type to sh 13118 1727204088.90332: variable 'ansible_shell_executable' from source: unknown 13118 1727204088.90337: variable 'ansible_connection' from source: unknown 13118 1727204088.90341: variable 'ansible_module_compression' from source: unknown 13118 1727204088.90343: variable 'ansible_shell_type' from source: unknown 13118 1727204088.90345: variable 'ansible_shell_executable' from source: unknown 13118 1727204088.90348: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204088.90352: variable 'ansible_pipelining' from source: unknown 13118 1727204088.90354: variable 'ansible_timeout' from source: unknown 13118 1727204088.90359: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204088.90496: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204088.90507: variable 'omit' from source: magic vars 13118 1727204088.90512: starting attempt loop 13118 1727204088.90515: running the handler 13118 1727204088.90531: _low_level_execute_command(): starting 13118 1727204088.90542: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204088.91286: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.91297: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.91307: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.91322: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.91363: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.91372: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.91382: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.91395: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.91403: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.91412: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.91418: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.91428: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.91443: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.91450: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.91458: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.91474: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.91550: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.91567: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.91573: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.91648: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.93198: stdout chunk (state=3): >>>/root <<< 13118 1727204088.93365: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204088.93368: stderr chunk (state=3): >>><<< 13118 1727204088.93373: stdout chunk (state=3): >>><<< 13118 1727204088.93399: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204088.93413: _low_level_execute_command(): starting 13118 1727204088.93418: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204088.933987-15322-201616282812233 `" && echo ansible-tmp-1727204088.933987-15322-201616282812233="` echo /root/.ansible/tmp/ansible-tmp-1727204088.933987-15322-201616282812233 `" ) && sleep 0' 13118 1727204088.94102: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.94111: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.94123: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.94139: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.94187: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.94195: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.94205: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.94219: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.94227: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.94245: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.94248: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.94254: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.94272: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.94281: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.94287: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.94300: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.94374: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.94387: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.94398: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.94653: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204088.96425: stdout chunk (state=3): >>>ansible-tmp-1727204088.933987-15322-201616282812233=/root/.ansible/tmp/ansible-tmp-1727204088.933987-15322-201616282812233 <<< 13118 1727204088.96571: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204088.96617: stderr chunk (state=3): >>><<< 13118 1727204088.96621: stdout chunk (state=3): >>><<< 13118 1727204088.96696: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204088.933987-15322-201616282812233=/root/.ansible/tmp/ansible-tmp-1727204088.933987-15322-201616282812233 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204088.96726: variable 'ansible_module_compression' from source: unknown 13118 1727204088.96788: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13118 1727204088.96827: variable 'ansible_facts' from source: unknown 13118 1727204088.97046: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204088.933987-15322-201616282812233/AnsiballZ_command.py 13118 1727204088.97185: Sending initial data 13118 1727204088.97188: Sent initial data (155 bytes) 13118 1727204088.98371: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204088.98387: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.98402: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.98422: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.98475: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.98484: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204088.98502: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.98518: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204088.98537: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204088.98547: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204088.98556: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204088.98567: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204088.98580: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204088.98588: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204088.98594: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204088.98604: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204088.98685: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204088.98703: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204088.98716: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204088.98800: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204089.00559: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204089.00567: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204089.00604: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmp8qvuyneo /root/.ansible/tmp/ansible-tmp-1727204088.933987-15322-201616282812233/AnsiballZ_command.py <<< 13118 1727204089.00651: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204089.01779: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204089.02002: stderr chunk (state=3): >>><<< 13118 1727204089.02006: stdout chunk (state=3): >>><<< 13118 1727204089.02009: done transferring module to remote 13118 1727204089.02015: _low_level_execute_command(): starting 13118 1727204089.02018: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204088.933987-15322-201616282812233/ /root/.ansible/tmp/ansible-tmp-1727204088.933987-15322-201616282812233/AnsiballZ_command.py && sleep 0' 13118 1727204089.02598: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204089.02613: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.02629: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.02651: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.02706: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204089.02718: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204089.02734: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.02763: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204089.02778: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204089.02792: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204089.02804: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.02818: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.02833: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.02854: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204089.02876: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204089.02910: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.02985: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204089.03008: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204089.03024: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204089.03090: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204089.04846: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204089.04928: stderr chunk (state=3): >>><<< 13118 1727204089.04931: stdout chunk (state=3): >>><<< 13118 1727204089.05055: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204089.05059: _low_level_execute_command(): starting 13118 1727204089.05062: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204088.933987-15322-201616282812233/AnsiballZ_command.py && sleep 0' 13118 1727204089.05745: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204089.05759: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.05783: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.05823: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.05875: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204089.05897: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204089.05918: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.05936: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204089.05947: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204089.05957: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204089.05971: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.05985: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.06022: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.06045: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204089.06057: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204089.06079: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.06174: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204089.06194: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204089.06210: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204089.06368: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204089.19949: stdout chunk (state=3): >>> {"changed": true, "stdout": "13: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000\n inet6 2001:db8::1df/128 scope global dynamic noprefixroute \n valid_lft 234sec preferred_lft 234sec\n inet6 2001:db8::af05:ee5:1cc5:17a4/64 scope global dynamic noprefixroute \n valid_lft 1796sec preferred_lft 1796sec\n inet6 fe80::cf66:9333:a14a:1e79/64 scope link noprefixroute \n valid_lft forever preferred_lft forever", "stderr": "", "rc": 0, "cmd": ["ip", "-6", "a", "s", "nm-bond"], "start": "2024-09-24 14:54:49.195122", "end": "2024-09-24 14:54:49.198764", "delta": "0:00:00.003642", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -6 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13118 1727204089.21217: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204089.21247: stderr chunk (state=3): >>><<< 13118 1727204089.21250: stdout chunk (state=3): >>><<< 13118 1727204089.21268: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "13: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000\n inet6 2001:db8::1df/128 scope global dynamic noprefixroute \n valid_lft 234sec preferred_lft 234sec\n inet6 2001:db8::af05:ee5:1cc5:17a4/64 scope global dynamic noprefixroute \n valid_lft 1796sec preferred_lft 1796sec\n inet6 fe80::cf66:9333:a14a:1e79/64 scope link noprefixroute \n valid_lft forever preferred_lft forever", "stderr": "", "rc": 0, "cmd": ["ip", "-6", "a", "s", "nm-bond"], "start": "2024-09-24 14:54:49.195122", "end": "2024-09-24 14:54:49.198764", "delta": "0:00:00.003642", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -6 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204089.21303: done with _execute_module (ansible.legacy.command, {'_raw_params': 'ip -6 a s nm-bond', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204088.933987-15322-201616282812233/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204089.21310: _low_level_execute_command(): starting 13118 1727204089.21315: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204088.933987-15322-201616282812233/ > /dev/null 2>&1 && sleep 0' 13118 1727204089.21763: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.21768: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.21799: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.21802: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.21805: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.21854: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204089.21857: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204089.21910: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204089.23740: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204089.23800: stderr chunk (state=3): >>><<< 13118 1727204089.23804: stdout chunk (state=3): >>><<< 13118 1727204089.23819: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204089.23825: handler run complete 13118 1727204089.23845: Evaluated conditional (False): False 13118 1727204089.23961: variable 'result' from source: set_fact 13118 1727204089.23982: Evaluated conditional ('2001' in result.stdout): True 13118 1727204089.23989: attempt loop complete, returning result 13118 1727204089.23992: _execute() done 13118 1727204089.23995: dumping result to json 13118 1727204089.24001: done dumping result, returning 13118 1727204089.24007: done running TaskExecutor() for managed-node2/TASK: ** TEST check IPv6 [0affcd87-79f5-56a3-0a64-000000000073] 13118 1727204089.24014: sending task result for task 0affcd87-79f5-56a3-0a64-000000000073 13118 1727204089.24112: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000073 13118 1727204089.24115: WORKER PROCESS EXITING ok: [managed-node2] => { "attempts": 1, "changed": false, "cmd": [ "ip", "-6", "a", "s", "nm-bond" ], "delta": "0:00:00.003642", "end": "2024-09-24 14:54:49.198764", "rc": 0, "start": "2024-09-24 14:54:49.195122" } STDOUT: 13: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000 inet6 2001:db8::1df/128 scope global dynamic noprefixroute valid_lft 234sec preferred_lft 234sec inet6 2001:db8::af05:ee5:1cc5:17a4/64 scope global dynamic noprefixroute valid_lft 1796sec preferred_lft 1796sec inet6 fe80::cf66:9333:a14a:1e79/64 scope link noprefixroute valid_lft forever preferred_lft forever 13118 1727204089.24195: no more pending results, returning what we have 13118 1727204089.24199: results queue empty 13118 1727204089.24200: checking for any_errors_fatal 13118 1727204089.24208: done checking for any_errors_fatal 13118 1727204089.24209: checking for max_fail_percentage 13118 1727204089.24210: done checking for max_fail_percentage 13118 1727204089.24211: checking to see if all hosts have failed and the running result is not ok 13118 1727204089.24212: done checking to see if all hosts have failed 13118 1727204089.24213: getting the remaining hosts for this loop 13118 1727204089.24214: done getting the remaining hosts for this loop 13118 1727204089.24218: getting the next task for host managed-node2 13118 1727204089.24233: done getting next task for host managed-node2 13118 1727204089.24237: ^ task is: TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role 13118 1727204089.24241: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204089.24257: getting variables 13118 1727204089.24258: in VariableManager get_vars() 13118 1727204089.24299: Calling all_inventory to load vars for managed-node2 13118 1727204089.24301: Calling groups_inventory to load vars for managed-node2 13118 1727204089.24303: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204089.24312: Calling all_plugins_play to load vars for managed-node2 13118 1727204089.24314: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204089.24317: Calling groups_plugins_play to load vars for managed-node2 13118 1727204089.25268: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204089.26192: done with get_vars() 13118 1727204089.26214: done getting variables TASK [fedora.linux_system_roles.network : Ensure ansible_facts used by role] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:4 Tuesday 24 September 2024 14:54:49 -0400 (0:00:00.375) 0:00:26.562 ***** 13118 1727204089.26289: entering _queue_task() for managed-node2/include_tasks 13118 1727204089.26525: worker is 1 (out of 1 available) 13118 1727204089.26541: exiting _queue_task() for managed-node2/include_tasks 13118 1727204089.26552: done queuing things up, now waiting for results queue to drain 13118 1727204089.26554: waiting for pending results... 13118 1727204089.26750: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role 13118 1727204089.26865: in run() - task 0affcd87-79f5-56a3-0a64-00000000007c 13118 1727204089.26878: variable 'ansible_search_path' from source: unknown 13118 1727204089.26881: variable 'ansible_search_path' from source: unknown 13118 1727204089.26911: calling self._execute() 13118 1727204089.26981: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204089.26993: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204089.26998: variable 'omit' from source: magic vars 13118 1727204089.27262: variable 'ansible_distribution_major_version' from source: facts 13118 1727204089.27276: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204089.27280: _execute() done 13118 1727204089.27284: dumping result to json 13118 1727204089.27292: done dumping result, returning 13118 1727204089.27298: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role [0affcd87-79f5-56a3-0a64-00000000007c] 13118 1727204089.27305: sending task result for task 0affcd87-79f5-56a3-0a64-00000000007c 13118 1727204089.27400: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000007c 13118 1727204089.27403: WORKER PROCESS EXITING 13118 1727204089.27469: no more pending results, returning what we have 13118 1727204089.27474: in VariableManager get_vars() 13118 1727204089.27521: Calling all_inventory to load vars for managed-node2 13118 1727204089.27524: Calling groups_inventory to load vars for managed-node2 13118 1727204089.27526: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204089.27539: Calling all_plugins_play to load vars for managed-node2 13118 1727204089.27541: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204089.27544: Calling groups_plugins_play to load vars for managed-node2 13118 1727204089.28347: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204089.29277: done with get_vars() 13118 1727204089.29291: variable 'ansible_search_path' from source: unknown 13118 1727204089.29292: variable 'ansible_search_path' from source: unknown 13118 1727204089.29319: we have included files to process 13118 1727204089.29320: generating all_blocks data 13118 1727204089.29321: done generating all_blocks data 13118 1727204089.29325: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13118 1727204089.29326: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13118 1727204089.29327: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13118 1727204089.29727: done processing included file 13118 1727204089.29729: iterating over new_blocks loaded from include file 13118 1727204089.29731: in VariableManager get_vars() 13118 1727204089.29749: done with get_vars() 13118 1727204089.29751: filtering new block on tags 13118 1727204089.29771: done filtering new block on tags 13118 1727204089.29773: in VariableManager get_vars() 13118 1727204089.29790: done with get_vars() 13118 1727204089.29791: filtering new block on tags 13118 1727204089.29817: done filtering new block on tags 13118 1727204089.29819: in VariableManager get_vars() 13118 1727204089.29835: done with get_vars() 13118 1727204089.29836: filtering new block on tags 13118 1727204089.29859: done filtering new block on tags 13118 1727204089.29860: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml for managed-node2 13118 1727204089.29866: extending task lists for all hosts with included blocks 13118 1727204089.30598: done extending task lists 13118 1727204089.30599: done processing included files 13118 1727204089.30600: results queue empty 13118 1727204089.30600: checking for any_errors_fatal 13118 1727204089.30603: done checking for any_errors_fatal 13118 1727204089.30604: checking for max_fail_percentage 13118 1727204089.30604: done checking for max_fail_percentage 13118 1727204089.30605: checking to see if all hosts have failed and the running result is not ok 13118 1727204089.30605: done checking to see if all hosts have failed 13118 1727204089.30606: getting the remaining hosts for this loop 13118 1727204089.30607: done getting the remaining hosts for this loop 13118 1727204089.30608: getting the next task for host managed-node2 13118 1727204089.30611: done getting next task for host managed-node2 13118 1727204089.30613: ^ task is: TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present 13118 1727204089.30615: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204089.30622: getting variables 13118 1727204089.30623: in VariableManager get_vars() 13118 1727204089.30634: Calling all_inventory to load vars for managed-node2 13118 1727204089.30636: Calling groups_inventory to load vars for managed-node2 13118 1727204089.30637: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204089.30641: Calling all_plugins_play to load vars for managed-node2 13118 1727204089.30642: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204089.30644: Calling groups_plugins_play to load vars for managed-node2 13118 1727204089.31317: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204089.32222: done with get_vars() 13118 1727204089.32239: done getting variables TASK [fedora.linux_system_roles.network : Ensure ansible_facts used by role are present] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:3 Tuesday 24 September 2024 14:54:49 -0400 (0:00:00.060) 0:00:26.622 ***** 13118 1727204089.32300: entering _queue_task() for managed-node2/setup 13118 1727204089.32637: worker is 1 (out of 1 available) 13118 1727204089.32650: exiting _queue_task() for managed-node2/setup 13118 1727204089.32665: done queuing things up, now waiting for results queue to drain 13118 1727204089.32667: waiting for pending results... 13118 1727204089.32963: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present 13118 1727204089.33137: in run() - task 0affcd87-79f5-56a3-0a64-000000000491 13118 1727204089.33155: variable 'ansible_search_path' from source: unknown 13118 1727204089.33161: variable 'ansible_search_path' from source: unknown 13118 1727204089.33202: calling self._execute() 13118 1727204089.33304: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204089.33315: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204089.33334: variable 'omit' from source: magic vars 13118 1727204089.33708: variable 'ansible_distribution_major_version' from source: facts 13118 1727204089.33727: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204089.33926: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204089.35483: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204089.35537: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204089.35590: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204089.35617: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204089.35642: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204089.35721: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204089.35750: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204089.35767: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204089.35898: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204089.35902: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204089.35904: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204089.35907: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204089.36075: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204089.36079: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204089.36082: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204089.36172: variable '__network_required_facts' from source: role '' defaults 13118 1727204089.36176: variable 'ansible_facts' from source: unknown 13118 1727204089.36938: Evaluated conditional (__network_required_facts | difference(ansible_facts.keys() | list) | length > 0): False 13118 1727204089.36941: when evaluation is False, skipping this task 13118 1727204089.36944: _execute() done 13118 1727204089.36947: dumping result to json 13118 1727204089.36949: done dumping result, returning 13118 1727204089.36957: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present [0affcd87-79f5-56a3-0a64-000000000491] 13118 1727204089.36963: sending task result for task 0affcd87-79f5-56a3-0a64-000000000491 13118 1727204089.37059: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000491 13118 1727204089.37062: WORKER PROCESS EXITING skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13118 1727204089.37103: no more pending results, returning what we have 13118 1727204089.37108: results queue empty 13118 1727204089.37109: checking for any_errors_fatal 13118 1727204089.37111: done checking for any_errors_fatal 13118 1727204089.37111: checking for max_fail_percentage 13118 1727204089.37113: done checking for max_fail_percentage 13118 1727204089.37114: checking to see if all hosts have failed and the running result is not ok 13118 1727204089.37114: done checking to see if all hosts have failed 13118 1727204089.37115: getting the remaining hosts for this loop 13118 1727204089.37116: done getting the remaining hosts for this loop 13118 1727204089.37120: getting the next task for host managed-node2 13118 1727204089.37133: done getting next task for host managed-node2 13118 1727204089.37137: ^ task is: TASK: fedora.linux_system_roles.network : Check if system is ostree 13118 1727204089.37142: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204089.37158: getting variables 13118 1727204089.37160: in VariableManager get_vars() 13118 1727204089.37203: Calling all_inventory to load vars for managed-node2 13118 1727204089.37206: Calling groups_inventory to load vars for managed-node2 13118 1727204089.37208: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204089.37218: Calling all_plugins_play to load vars for managed-node2 13118 1727204089.37220: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204089.37222: Calling groups_plugins_play to load vars for managed-node2 13118 1727204089.38780: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204089.40532: done with get_vars() 13118 1727204089.40558: done getting variables TASK [fedora.linux_system_roles.network : Check if system is ostree] *********** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:12 Tuesday 24 September 2024 14:54:49 -0400 (0:00:00.083) 0:00:26.706 ***** 13118 1727204089.40691: entering _queue_task() for managed-node2/stat 13118 1727204089.41157: worker is 1 (out of 1 available) 13118 1727204089.41173: exiting _queue_task() for managed-node2/stat 13118 1727204089.41188: done queuing things up, now waiting for results queue to drain 13118 1727204089.41189: waiting for pending results... 13118 1727204089.41519: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check if system is ostree 13118 1727204089.41753: in run() - task 0affcd87-79f5-56a3-0a64-000000000493 13118 1727204089.41785: variable 'ansible_search_path' from source: unknown 13118 1727204089.41793: variable 'ansible_search_path' from source: unknown 13118 1727204089.41870: calling self._execute() 13118 1727204089.41992: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204089.42019: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204089.42546: variable 'omit' from source: magic vars 13118 1727204089.42939: variable 'ansible_distribution_major_version' from source: facts 13118 1727204089.42956: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204089.43140: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204089.43426: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204089.43480: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204089.43525: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204089.43569: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204089.43663: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204089.43696: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204089.43735: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204089.43768: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204089.43872: variable '__network_is_ostree' from source: set_fact 13118 1727204089.43884: Evaluated conditional (not __network_is_ostree is defined): False 13118 1727204089.43891: when evaluation is False, skipping this task 13118 1727204089.43897: _execute() done 13118 1727204089.43904: dumping result to json 13118 1727204089.43911: done dumping result, returning 13118 1727204089.43922: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check if system is ostree [0affcd87-79f5-56a3-0a64-000000000493] 13118 1727204089.43935: sending task result for task 0affcd87-79f5-56a3-0a64-000000000493 skipping: [managed-node2] => { "changed": false, "false_condition": "not __network_is_ostree is defined", "skip_reason": "Conditional result was False" } 13118 1727204089.44112: no more pending results, returning what we have 13118 1727204089.44117: results queue empty 13118 1727204089.44118: checking for any_errors_fatal 13118 1727204089.44125: done checking for any_errors_fatal 13118 1727204089.44126: checking for max_fail_percentage 13118 1727204089.44127: done checking for max_fail_percentage 13118 1727204089.44128: checking to see if all hosts have failed and the running result is not ok 13118 1727204089.44129: done checking to see if all hosts have failed 13118 1727204089.44133: getting the remaining hosts for this loop 13118 1727204089.44134: done getting the remaining hosts for this loop 13118 1727204089.44139: getting the next task for host managed-node2 13118 1727204089.44148: done getting next task for host managed-node2 13118 1727204089.44151: ^ task is: TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree 13118 1727204089.44156: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204089.44176: getting variables 13118 1727204089.44178: in VariableManager get_vars() 13118 1727204089.44222: Calling all_inventory to load vars for managed-node2 13118 1727204089.44225: Calling groups_inventory to load vars for managed-node2 13118 1727204089.44228: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204089.44241: Calling all_plugins_play to load vars for managed-node2 13118 1727204089.44244: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204089.44247: Calling groups_plugins_play to load vars for managed-node2 13118 1727204089.45328: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000493 13118 1727204089.45336: WORKER PROCESS EXITING 13118 1727204089.46271: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204089.50228: done with get_vars() 13118 1727204089.50270: done getting variables 13118 1727204089.50327: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Set flag to indicate system is ostree] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:17 Tuesday 24 September 2024 14:54:49 -0400 (0:00:00.096) 0:00:26.803 ***** 13118 1727204089.50369: entering _queue_task() for managed-node2/set_fact 13118 1727204089.50697: worker is 1 (out of 1 available) 13118 1727204089.50709: exiting _queue_task() for managed-node2/set_fact 13118 1727204089.50720: done queuing things up, now waiting for results queue to drain 13118 1727204089.50722: waiting for pending results... 13118 1727204089.51336: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree 13118 1727204089.51790: in run() - task 0affcd87-79f5-56a3-0a64-000000000494 13118 1727204089.51802: variable 'ansible_search_path' from source: unknown 13118 1727204089.51807: variable 'ansible_search_path' from source: unknown 13118 1727204089.51840: calling self._execute() 13118 1727204089.51932: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204089.51936: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204089.51945: variable 'omit' from source: magic vars 13118 1727204089.52294: variable 'ansible_distribution_major_version' from source: facts 13118 1727204089.52307: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204089.53280: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204089.53766: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204089.53814: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204089.53846: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204089.54085: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204089.54374: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204089.54401: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204089.54428: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204089.54453: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204089.54548: variable '__network_is_ostree' from source: set_fact 13118 1727204089.54556: Evaluated conditional (not __network_is_ostree is defined): False 13118 1727204089.54559: when evaluation is False, skipping this task 13118 1727204089.54562: _execute() done 13118 1727204089.54565: dumping result to json 13118 1727204089.54570: done dumping result, returning 13118 1727204089.54583: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree [0affcd87-79f5-56a3-0a64-000000000494] 13118 1727204089.54585: sending task result for task 0affcd87-79f5-56a3-0a64-000000000494 13118 1727204089.54689: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000494 13118 1727204089.54692: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "not __network_is_ostree is defined", "skip_reason": "Conditional result was False" } 13118 1727204089.54746: no more pending results, returning what we have 13118 1727204089.54751: results queue empty 13118 1727204089.54752: checking for any_errors_fatal 13118 1727204089.54760: done checking for any_errors_fatal 13118 1727204089.54760: checking for max_fail_percentage 13118 1727204089.54763: done checking for max_fail_percentage 13118 1727204089.54765: checking to see if all hosts have failed and the running result is not ok 13118 1727204089.54767: done checking to see if all hosts have failed 13118 1727204089.54767: getting the remaining hosts for this loop 13118 1727204089.54769: done getting the remaining hosts for this loop 13118 1727204089.54773: getting the next task for host managed-node2 13118 1727204089.54785: done getting next task for host managed-node2 13118 1727204089.54789: ^ task is: TASK: fedora.linux_system_roles.network : Check which services are running 13118 1727204089.54795: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204089.54817: getting variables 13118 1727204089.54819: in VariableManager get_vars() 13118 1727204089.54869: Calling all_inventory to load vars for managed-node2 13118 1727204089.54872: Calling groups_inventory to load vars for managed-node2 13118 1727204089.54875: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204089.54886: Calling all_plugins_play to load vars for managed-node2 13118 1727204089.54890: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204089.54893: Calling groups_plugins_play to load vars for managed-node2 13118 1727204089.57480: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204089.59336: done with get_vars() 13118 1727204089.59410: done getting variables TASK [fedora.linux_system_roles.network : Check which services are running] **** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:21 Tuesday 24 September 2024 14:54:49 -0400 (0:00:00.092) 0:00:26.895 ***** 13118 1727204089.59578: entering _queue_task() for managed-node2/service_facts 13118 1727204089.59981: worker is 1 (out of 1 available) 13118 1727204089.59996: exiting _queue_task() for managed-node2/service_facts 13118 1727204089.60009: done queuing things up, now waiting for results queue to drain 13118 1727204089.60015: waiting for pending results... 13118 1727204089.60316: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check which services are running 13118 1727204089.60511: in run() - task 0affcd87-79f5-56a3-0a64-000000000496 13118 1727204089.60533: variable 'ansible_search_path' from source: unknown 13118 1727204089.60540: variable 'ansible_search_path' from source: unknown 13118 1727204089.60586: calling self._execute() 13118 1727204089.60698: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204089.60710: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204089.60725: variable 'omit' from source: magic vars 13118 1727204089.61113: variable 'ansible_distribution_major_version' from source: facts 13118 1727204089.61134: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204089.61147: variable 'omit' from source: magic vars 13118 1727204089.61244: variable 'omit' from source: magic vars 13118 1727204089.61287: variable 'omit' from source: magic vars 13118 1727204089.61344: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204089.61422: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204089.61459: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204089.61484: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204089.61562: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204089.61596: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204089.61774: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204089.61781: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204089.61878: Set connection var ansible_timeout to 10 13118 1727204089.61893: Set connection var ansible_pipelining to False 13118 1727204089.61898: Set connection var ansible_connection to ssh 13118 1727204089.61905: Set connection var ansible_shell_executable to /bin/sh 13118 1727204089.61912: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204089.61917: Set connection var ansible_shell_type to sh 13118 1727204089.61949: variable 'ansible_shell_executable' from source: unknown 13118 1727204089.61957: variable 'ansible_connection' from source: unknown 13118 1727204089.61968: variable 'ansible_module_compression' from source: unknown 13118 1727204089.61980: variable 'ansible_shell_type' from source: unknown 13118 1727204089.61986: variable 'ansible_shell_executable' from source: unknown 13118 1727204089.61992: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204089.62018: variable 'ansible_pipelining' from source: unknown 13118 1727204089.62025: variable 'ansible_timeout' from source: unknown 13118 1727204089.62037: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204089.62260: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204089.62277: variable 'omit' from source: magic vars 13118 1727204089.62286: starting attempt loop 13118 1727204089.62293: running the handler 13118 1727204089.62319: _low_level_execute_command(): starting 13118 1727204089.62334: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204089.63405: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204089.63427: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.63448: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.63472: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.63515: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204089.63532: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204089.63554: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.63576: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204089.63589: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204089.63601: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204089.63615: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.63633: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.63658: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.63676: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204089.63689: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204089.63705: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.63803: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204089.63826: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204089.63848: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204089.64034: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204089.65618: stdout chunk (state=3): >>>/root <<< 13118 1727204089.65838: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204089.65842: stdout chunk (state=3): >>><<< 13118 1727204089.65844: stderr chunk (state=3): >>><<< 13118 1727204089.65975: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204089.65979: _low_level_execute_command(): starting 13118 1727204089.65981: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204089.658723-15364-162876841556701 `" && echo ansible-tmp-1727204089.658723-15364-162876841556701="` echo /root/.ansible/tmp/ansible-tmp-1727204089.658723-15364-162876841556701 `" ) && sleep 0' 13118 1727204089.66674: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.66677: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.66711: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204089.66714: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.66718: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.66784: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204089.67335: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204089.67388: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204089.69262: stdout chunk (state=3): >>>ansible-tmp-1727204089.658723-15364-162876841556701=/root/.ansible/tmp/ansible-tmp-1727204089.658723-15364-162876841556701 <<< 13118 1727204089.69451: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204089.69455: stdout chunk (state=3): >>><<< 13118 1727204089.69507: stderr chunk (state=3): >>><<< 13118 1727204089.69510: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204089.658723-15364-162876841556701=/root/.ansible/tmp/ansible-tmp-1727204089.658723-15364-162876841556701 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204089.69553: variable 'ansible_module_compression' from source: unknown 13118 1727204089.69871: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.service_facts-ZIP_DEFLATED 13118 1727204089.69875: variable 'ansible_facts' from source: unknown 13118 1727204089.69877: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204089.658723-15364-162876841556701/AnsiballZ_service_facts.py 13118 1727204089.69879: Sending initial data 13118 1727204089.69881: Sent initial data (161 bytes) 13118 1727204089.70849: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204089.70859: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.70872: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.70911: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.71101: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204089.71105: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204089.71107: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.71110: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204089.71112: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204089.71114: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204089.71116: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.71118: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.71120: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.71126: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204089.71128: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204089.71132: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.71135: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204089.71140: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204089.71304: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204089.71307: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204089.72942: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204089.72983: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204089.73023: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpbamnp53x /root/.ansible/tmp/ansible-tmp-1727204089.658723-15364-162876841556701/AnsiballZ_service_facts.py <<< 13118 1727204089.73053: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204089.74400: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204089.74643: stderr chunk (state=3): >>><<< 13118 1727204089.74647: stdout chunk (state=3): >>><<< 13118 1727204089.74649: done transferring module to remote 13118 1727204089.74651: _low_level_execute_command(): starting 13118 1727204089.74654: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204089.658723-15364-162876841556701/ /root/.ansible/tmp/ansible-tmp-1727204089.658723-15364-162876841556701/AnsiballZ_service_facts.py && sleep 0' 13118 1727204089.75315: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204089.75329: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.75346: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.75375: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.75424: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204089.75441: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204089.75456: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.75477: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204089.75489: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204089.75500: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204089.75516: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.75533: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.75550: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.75562: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204089.75577: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204089.75591: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.75692: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204089.75720: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204089.75749: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204089.75827: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204089.77684: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204089.77687: stdout chunk (state=3): >>><<< 13118 1727204089.77690: stderr chunk (state=3): >>><<< 13118 1727204089.77784: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204089.77788: _low_level_execute_command(): starting 13118 1727204089.77790: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204089.658723-15364-162876841556701/AnsiballZ_service_facts.py && sleep 0' 13118 1727204089.78358: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204089.78374: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.78387: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.78404: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.78448: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204089.78460: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204089.78482: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.78499: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204089.78510: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204089.78520: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204089.78533: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204089.78546: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204089.78561: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204089.78576: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204089.78586: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204089.78598: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204089.78681: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204089.78698: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204089.78712: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204089.78791: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204091.07747: stdout chunk (state=3): >>> {"ansible_facts": {"services": {"auditd.service": {"name": "auditd.service", "state": "running", "status": "enabled", "source": "systemd"}, "auth-rpcgss-module.service": {"name": "auth-rpcgss-module.service", "state": "stopped", "status": "static", "source": "systemd"}, "autofs.service": {"name": "autofs.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "chronyd.service": {"name": "chronyd.service", "state": "running", "status": "enabled", "source": "systemd"}, "cloud-config.service": {"name": "cloud-config.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-final.service": {"name": "cloud-final.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init-local.service": {"name": "cloud-init-local.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init.service": {"name": "cloud-init.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "crond.service": {"name": "crond.service", "state": "running", "status": "enabled", "source": "systemd"}, "dbus-broker.service": {"name": "dbus-broker.service", "state": "running", "status": "enabled", "source": "systemd"}, "display-manager.service": {"name": "display-manager.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "dnf-makecache.service": {"name": "dnf-makecache.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-cmdline.service": {"name": "dracut-cmdline.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-initqueue.service": {"name": "dracut-initqueue.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-mount.service": {"name": "dracut-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-mount.service": {"name": "dracut-pre-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-pivot.service": {"name": "dracut-pre-pivot.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-trigger.service": {"name": "dracut-pre-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-udev.service": {"name": "dracut-pre-udev.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown-onfailure.service": {"name": "dracut-shutdown-onfailure.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown.service": {"name": "dracut-shutdown.service", "state": "stopped", "status": "static", "source": "systemd"}, "emergency.service": {"name": "emergency.service", "state": "stopped", "status": "static", "source": "systemd"}, "getty@tty1.service": {"name": "getty@tty1.service", "state": "running", "status": "active", "source": "systemd"}, "gssproxy.service": {"name": "gssproxy.service", "state": "running", "status": "disabled", "source": "systemd"}, "hv_kvp_daemon.service": {"name": "hv_kvp_daemon.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "initrd-cleanup.service": {"name": "initrd-cleanup.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-parse-etc.service": {"name": "initrd-parse-etc.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-switch-root.service": {"name": "initrd-switch-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-udevadm-cleanup-db.service": {"name": "initrd-udevadm-cleanup-db.service", "state": "stopped", "status": "static", "source": "systemd"}, "irqbalance.service": {"name": "irqbalance.service", "state": "running", "status": "enabled", "source": "systemd"}, "kdump.service": {"name": "kdump.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "kmod-static-nodes.service": {"name": "kmod-static-nodes.service", "state": "stopped", "status": "static", "source": "systemd"}, "ldconfig.service": {"name": "ldconfig.service", "state": "stopped", "status": "static", "source": "systemd"}, "logrotate.service": {"name": "logrotate.service", "state": "stopped", "status": "static", "source": "systemd"}, "microcode.service": {"name": "microcode.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "modprobe@configfs.service": {"name": "modprobe@configfs.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@drm.service": {"name": "modprobe@drm.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@fuse.service": {"name": "modprobe@fuse.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "network.service": {"name": "network.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "NetworkManager-dispatcher.service": {"name": "NetworkManager-dispatcher.service", "state": "running", "status": "enabled", "source": "systemd"}, "NetworkManager-wait-online.service": {"name": "NetworkManager-wait-online.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "NetworkManager.service": {"name": "NetworkManager.service", "state": "running", "status": "enabled", "source": "systemd"}, "nfs-idmapd.service": {"name": "nfs-idmapd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-mountd.service": {"name": "nfs-mountd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-server.service": {"name": "nfs-server.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "nfs-utils.service": {"name": "nfs-utils.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfsdcld.service": {"name": "nfsdcld.service", "state": "stopped", "status": "static", "source": "systemd"}, "nis-domainname.service": {"name": "nis-domainname.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "ntpd.service": {"name": "ntpd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ntpdate.service": {"name": "ntpdate.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-quit-wait.service": {"name": "plymouth-qu<<< 13118 1727204091.07769: stdout chunk (state=3): >>>it-wait.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-start.service": {"name": "plymouth-start.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rc-local.service": {"name": "rc-local.service", "state": "stopped", "status": "static", "source": "systemd"}, "rescue.service": {"name": "rescue.service", "state": "stopped", "status": "static", "source": "systemd"}, "restraintd.service": {"name": "restraintd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rngd.service": {"name": "rngd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rpc-gssd.service": {"name": "rpc-gssd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd-notify.service": {"name": "rpc-statd-notify.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd.service": {"name": "rpc-statd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-svcgssd.service": {"name": "rpc-svcgssd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rpcbind.service": {"name": "rpcbind.service", "state": "running", "status": "enabled", "source": "systemd"}, "rsyslog.service": {"name": "rsyslog.service", "state": "running", "status": "enabled", "source": "systemd"}, "selinux-autorelabel-mark.service": {"name": "selinux-autorelabel-mark.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "serial-getty@ttyS0.service": {"name": "serial-getty@ttyS0.service", "state": "running", "status": "active", "source": "systemd"}, "snapd.seeded.service": {"name": "snapd.seeded.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sntp.service": {"name": "sntp.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen.service": {"name": "sshd-keygen.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen@ecdsa.service": {"name": "sshd-keygen@ecdsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@ed25519.service": {"name": "sshd-keygen@ed25519.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@rsa.service": {"name": "sshd-keygen@rsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd.service": {"name": "sshd.service", "state": "running", "status": "enabled", "source": "systemd"}, "sssd-kcm.service": {"name": "sssd-kcm.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "sssd.service": {"name": "sssd.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "syslog.service": {"name": "syslog.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-ask-password-console.service": {"name": "systemd-ask-password-console.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-ask-password-wall.service": {"name": "systemd-ask-password-wall.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-binfmt.service": {"name": "systemd-binfmt.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-random-seed.service": {"name": "systemd-boot-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-update.service": {"name": "systemd-boot-update.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-firstboot.service": {"name": "systemd-firstboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-fsck-root.service": {"name": "systemd-fsck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hwdb-update.service": {"name": "systemd-hwdb-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-initctl.service": {"name": "systemd-initctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-catalog-update.service": {"name": "systemd-journal-catalog-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-flush.service": {"name": "systemd-journal-flush.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journald.service": {"name": "systemd-journald.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-logind.service": {"name": "systemd-logind.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-machine-id-commit.service": {"name": "systemd-machine-id-commit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-modules-load.service": {"name": "systemd-modules-load.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-network-generator.service": {"name": "systemd-network-generator.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-networkd-wait-online.service": {"name": "systemd-networkd-wait-online.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-pcrmachine.service": {"name": "systemd-pcrmachine.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-initrd.service": {"name": "systemd-pcrphase-initrd.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-sysinit.service": {"name": "systemd-pcrphase-sysinit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase.service": {"name": "systemd-pcrphase.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-random-seed.service": {"name": "systemd-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-remount-fs.service": {"name": "systemd-remount-fs.service", "state": "stopped", "status": "enabled-runtime", "source": "systemd"}, "systemd-repart.service": {"name": "systemd-repart.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-rfkill.service": {"name": "systemd-rfkill.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysctl.service": {"name": "systemd-sysctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysext.service": {"name": "systemd-sysext.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "systemd-sysusers.service": {"name": "systemd-sysusers.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-timesyncd.service": {"name": "systemd-timesyncd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-tmpfiles-clean.service": {"name": "systemd-tmpfiles-clean.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev.service": {"name": "systemd-tmpfiles-setup-dev.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup.service": {"name": "systemd-tmpfiles-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles.service": {"name": "systemd-tmpfiles.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-udev-settle.service": {"name": "systemd-udev-settle.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-trigger.service": {"name": "systemd-udev-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udevd.service": {"name": "systemd-udevd.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-update-done.service": {"name": "systemd-update-done.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp-runlevel.service": {"name": "systemd-update-utmp-runlevel.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp.service": {"name": "systemd-update-utmp.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-user-sessions.service": {"name": "systemd-user-sessions.service", "state": "stopped", "status": "static", "source": "systemd"}, "syst<<< 13118 1727204091.07795: stdout chunk (state=3): >>>emd-vconsole-setup.service": {"name": "systemd-vconsole-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "user-runtime-dir@0.service": {"name": "user-runtime-dir@0.service", "state": "stopped", "status": "active", "source": "systemd"}, "user@0.service": {"name": "user@0.service", "state": "running", "status": "active", "source": "systemd"}, "wpa_supplicant.service": {"name": "wpa_supplicant.service", "state": "running", "status": "enabled", "source": "systemd"}, "ypbind.service": {"name": "ypbind.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "yppasswdd.service": {"name": "yppasswdd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ypserv.service": {"name": "ypserv.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ypxfrd.service": {"name": "ypxfrd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "autovt@.service": {"name": "autovt@.service", "state": "unknown", "status": "alias", "source": "systemd"}, "chrony-wait.service": {"name": "chrony-wait.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "chronyd-restricted.service": {"name": "chronyd-restricted.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "cloud-init-hotplugd.service": {"name": "cloud-init-hotplugd.service", "state": "inactive", "status": "static", "source": "systemd"}, "console-getty.service": {"name": "console-getty.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "container-getty@.service": {"name": "container-getty@.service", "state": "unknown", "status": "static", "source": "systemd"}, "cpupower.service": {"name": "cpupower.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dbus-org.freedesktop.hostname1.service": {"name": "dbus-org.freedesktop.hostname1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.locale1.service": {"name": "dbus-org.freedesktop.locale1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.login1.service": {"name": "dbus-org.freedesktop.login1.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.nm-dispatcher.service": {"name": "dbus-org.freedesktop.nm-dispatcher.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.timedate1.service": {"name": "dbus-org.freedesktop.timedate1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus.service": {"name": "dbus.service", "state": "active", "status": "alias", "source": "systemd"}, "debug-shell.service": {"name": "debug-shell.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnf-system-upgrade-cleanup.service": {"name": "dnf-system-upgrade-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "dnf-system-upgrade.service": {"name": "dnf-system-upgrade.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnsmasq.service": {"name": "dnsmasq.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "firewalld.service": {"name": "firewalld.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fstrim.service": {"name": "fstrim.service", "state": "inactive", "status": "static", "source": "systemd"}, "getty@.service": {"name": "getty@.service", "state": "unknown", "status": "enabled", "source": "systemd"}, "grub-boot-indeterminate.service": {"name": "grub-boot-indeterminate.service", "state": "inactive", "status": "static", "source": "systemd"}, "grub2-systemd-integration.service": {"name": "grub2-systemd-integration.service", "state": "inactive", "status": "static", "source": "systemd"}, "hostapd.service": {"name": "hostapd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "kvm_stat.service": {"name": "kvm_stat.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "man-db-cache-update.service": {"name": "man-db-cache-update.service", "state": "inactive", "status": "static", "source": "systemd"}, "man-db-restart-cache-update.service": {"name": "man-db-restart-cache-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "modprobe@.service": {"name": "modprobe@.service", "state": "unknown", "status": "static", "source": "systemd"}, "nfs-blkmap.service": {"name": "nfs-blkmap.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nftables.service": {"name": "nftables.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nm-priv-helper.service": {"name": "nm-priv-helper.service", "state": "inactive", "status": "static", "source": "systemd"}, "oddjobd.service": {"name": "oddjobd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "pam_namespace.service": {"name": "pam_namespace.service", "state": "inactive", "status": "static", "source": "systemd"}, "qemu-guest-agent.service": {"name": "qemu-guest-agent.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "quotaon.service": {"name": "quotaon.service", "state": "inactive", "status": "static", "source": "systemd"}, "rdisc.service": {"name": "rdisc.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "rpmdb-rebuild.service": {"name": "rpmdb-rebuild.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "selinux-autorelabel.service": {"name": "selinux-autorelabel.service", "state": "inactive", "status": "static", "source": "systemd"}, "selinux-check-proper-disable.service": {"name": "selinux-check-proper-disable.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "serial-getty@.service": {"name": "serial-getty@.service", "state": "unknown", "status": "indirect", "source": "systemd"}, "sshd-keygen@.service": {"name": "sshd-keygen@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "sshd@.service": {"name": "sshd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "sssd-autofs.service": {"name": "sssd-autofs.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-nss.service": {"name": "sssd-nss.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pac.service": {"name": "sssd-pac.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pam.service": {"name": "sssd-pam.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-ssh.service": {"name": "sssd-ssh.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-sudo.service": {"name": "sssd-sudo.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "system-update-cleanup.service": {"name": "system-update-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-backlight@.service": {"name": "systemd-backlight@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-bless-boot.service": {"name": "systemd-bless-boot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-boot-check-no-failures.service": {"name": "systemd-boot-check-no-failures.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-coredump@.service": {"name": "systemd-coredump@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-exit.service": {"name": "systemd-exit.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-fsck@.service": {"name": "systemd-fsck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-growfs-root.service": {"name": "systemd-growfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-growfs@.service": {"name": "systemd-growfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-halt.service": {"name": "systemd-halt.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hibernate-resume@.service": {"name": "systemd-hibernate-resume@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-hibernate.service": {"name": "systemd-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hostnamed.service": {"name": "systemd-hostnamed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hybrid-sleep.service": {"name": "systemd-hybrid-sleep.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-journald@.service": {"name": "systemd-journald@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-kexec.service": {"name": "systemd-kexec.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-localed.service": {"name": "systemd-localed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs-root.service": {"name": "systemd-pcrfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs@.service": {"name": "systemd-pcrfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-poweroff.service": {"name": "systemd-poweroff.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pstore.service": {"name": "systemd-pstore.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-quotacheck.service": {"name": "systemd-quotacheck.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-reboot.service": {"name": "systemd-reboot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend-then-hibernate.service": {"name": "systemd-suspend-then-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend.service": {"name": "systemd-suspend.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-sysupdate-reboot.service": {"name": "systemd-sysupdate-reboot.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-sysupdate.service": {"name": "systemd-sysupdate.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-timedated.service": {"name": "systemd-timedated.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-volatile-root.service": {"name": "systemd-volatile-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "teamd@.service": {"name": "teamd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user-runtime-dir@.service": {"name": "user-runtime-dir@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user@.service": {"name": "user@.service", "state": "unknown", "status": "static", "source": "systemd"}}}, "invocation": {"module_args": {}}} <<< 13118 1727204091.09199: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204091.09246: stderr chunk (state=3): >>><<< 13118 1727204091.09249: stdout chunk (state=3): >>><<< 13118 1727204091.09476: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"services": {"auditd.service": {"name": "auditd.service", "state": "running", "status": "enabled", "source": "systemd"}, "auth-rpcgss-module.service": {"name": "auth-rpcgss-module.service", "state": "stopped", "status": "static", "source": "systemd"}, "autofs.service": {"name": "autofs.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "chronyd.service": {"name": "chronyd.service", "state": "running", "status": "enabled", "source": "systemd"}, "cloud-config.service": {"name": "cloud-config.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-final.service": {"name": "cloud-final.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init-local.service": {"name": "cloud-init-local.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init.service": {"name": "cloud-init.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "crond.service": {"name": "crond.service", "state": "running", "status": "enabled", "source": "systemd"}, "dbus-broker.service": {"name": "dbus-broker.service", "state": "running", "status": "enabled", "source": "systemd"}, "display-manager.service": {"name": "display-manager.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "dnf-makecache.service": {"name": "dnf-makecache.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-cmdline.service": {"name": "dracut-cmdline.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-initqueue.service": {"name": "dracut-initqueue.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-mount.service": {"name": "dracut-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-mount.service": {"name": "dracut-pre-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-pivot.service": {"name": "dracut-pre-pivot.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-trigger.service": {"name": "dracut-pre-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-udev.service": {"name": "dracut-pre-udev.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown-onfailure.service": {"name": "dracut-shutdown-onfailure.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown.service": {"name": "dracut-shutdown.service", "state": "stopped", "status": "static", "source": "systemd"}, "emergency.service": {"name": "emergency.service", "state": "stopped", "status": "static", "source": "systemd"}, "getty@tty1.service": {"name": "getty@tty1.service", "state": "running", "status": "active", "source": "systemd"}, "gssproxy.service": {"name": "gssproxy.service", "state": "running", "status": "disabled", "source": "systemd"}, "hv_kvp_daemon.service": {"name": "hv_kvp_daemon.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "initrd-cleanup.service": {"name": "initrd-cleanup.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-parse-etc.service": {"name": "initrd-parse-etc.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-switch-root.service": {"name": "initrd-switch-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-udevadm-cleanup-db.service": {"name": "initrd-udevadm-cleanup-db.service", "state": "stopped", "status": "static", "source": "systemd"}, "irqbalance.service": {"name": "irqbalance.service", "state": "running", "status": "enabled", "source": "systemd"}, "kdump.service": {"name": "kdump.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "kmod-static-nodes.service": {"name": "kmod-static-nodes.service", "state": "stopped", "status": "static", "source": "systemd"}, "ldconfig.service": {"name": "ldconfig.service", "state": "stopped", "status": "static", "source": "systemd"}, "logrotate.service": {"name": "logrotate.service", "state": "stopped", "status": "static", "source": "systemd"}, "microcode.service": {"name": "microcode.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "modprobe@configfs.service": {"name": "modprobe@configfs.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@drm.service": {"name": "modprobe@drm.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@fuse.service": {"name": "modprobe@fuse.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "network.service": {"name": "network.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "NetworkManager-dispatcher.service": {"name": "NetworkManager-dispatcher.service", "state": "running", "status": "enabled", "source": "systemd"}, "NetworkManager-wait-online.service": {"name": "NetworkManager-wait-online.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "NetworkManager.service": {"name": "NetworkManager.service", "state": "running", "status": "enabled", "source": "systemd"}, "nfs-idmapd.service": {"name": "nfs-idmapd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-mountd.service": {"name": "nfs-mountd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-server.service": {"name": "nfs-server.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "nfs-utils.service": {"name": "nfs-utils.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfsdcld.service": {"name": "nfsdcld.service", "state": "stopped", "status": "static", "source": "systemd"}, "nis-domainname.service": {"name": "nis-domainname.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "ntpd.service": {"name": "ntpd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ntpdate.service": {"name": "ntpdate.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-quit-wait.service": {"name": "plymouth-quit-wait.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-start.service": {"name": "plymouth-start.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rc-local.service": {"name": "rc-local.service", "state": "stopped", "status": "static", "source": "systemd"}, "rescue.service": {"name": "rescue.service", "state": "stopped", "status": "static", "source": "systemd"}, "restraintd.service": {"name": "restraintd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rngd.service": {"name": "rngd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rpc-gssd.service": {"name": "rpc-gssd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd-notify.service": {"name": "rpc-statd-notify.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd.service": {"name": "rpc-statd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-svcgssd.service": {"name": "rpc-svcgssd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rpcbind.service": {"name": "rpcbind.service", "state": "running", "status": "enabled", "source": "systemd"}, "rsyslog.service": {"name": "rsyslog.service", "state": "running", "status": "enabled", "source": "systemd"}, "selinux-autorelabel-mark.service": {"name": "selinux-autorelabel-mark.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "serial-getty@ttyS0.service": {"name": "serial-getty@ttyS0.service", "state": "running", "status": "active", "source": "systemd"}, "snapd.seeded.service": {"name": "snapd.seeded.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sntp.service": {"name": "sntp.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen.service": {"name": "sshd-keygen.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen@ecdsa.service": {"name": "sshd-keygen@ecdsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@ed25519.service": {"name": "sshd-keygen@ed25519.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@rsa.service": {"name": "sshd-keygen@rsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd.service": {"name": "sshd.service", "state": "running", "status": "enabled", "source": "systemd"}, "sssd-kcm.service": {"name": "sssd-kcm.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "sssd.service": {"name": "sssd.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "syslog.service": {"name": "syslog.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-ask-password-console.service": {"name": "systemd-ask-password-console.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-ask-password-wall.service": {"name": "systemd-ask-password-wall.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-binfmt.service": {"name": "systemd-binfmt.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-random-seed.service": {"name": "systemd-boot-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-update.service": {"name": "systemd-boot-update.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-firstboot.service": {"name": "systemd-firstboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-fsck-root.service": {"name": "systemd-fsck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hwdb-update.service": {"name": "systemd-hwdb-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-initctl.service": {"name": "systemd-initctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-catalog-update.service": {"name": "systemd-journal-catalog-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-flush.service": {"name": "systemd-journal-flush.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journald.service": {"name": "systemd-journald.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-logind.service": {"name": "systemd-logind.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-machine-id-commit.service": {"name": "systemd-machine-id-commit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-modules-load.service": {"name": "systemd-modules-load.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-network-generator.service": {"name": "systemd-network-generator.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-networkd-wait-online.service": {"name": "systemd-networkd-wait-online.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-pcrmachine.service": {"name": "systemd-pcrmachine.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-initrd.service": {"name": "systemd-pcrphase-initrd.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-sysinit.service": {"name": "systemd-pcrphase-sysinit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase.service": {"name": "systemd-pcrphase.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-random-seed.service": {"name": "systemd-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-remount-fs.service": {"name": "systemd-remount-fs.service", "state": "stopped", "status": "enabled-runtime", "source": "systemd"}, "systemd-repart.service": {"name": "systemd-repart.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-rfkill.service": {"name": "systemd-rfkill.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysctl.service": {"name": "systemd-sysctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysext.service": {"name": "systemd-sysext.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "systemd-sysusers.service": {"name": "systemd-sysusers.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-timesyncd.service": {"name": "systemd-timesyncd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-tmpfiles-clean.service": {"name": "systemd-tmpfiles-clean.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev.service": {"name": "systemd-tmpfiles-setup-dev.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup.service": {"name": "systemd-tmpfiles-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles.service": {"name": "systemd-tmpfiles.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-udev-settle.service": {"name": "systemd-udev-settle.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-trigger.service": {"name": "systemd-udev-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udevd.service": {"name": "systemd-udevd.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-update-done.service": {"name": "systemd-update-done.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp-runlevel.service": {"name": "systemd-update-utmp-runlevel.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp.service": {"name": "systemd-update-utmp.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-user-sessions.service": {"name": "systemd-user-sessions.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-vconsole-setup.service": {"name": "systemd-vconsole-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "user-runtime-dir@0.service": {"name": "user-runtime-dir@0.service", "state": "stopped", "status": "active", "source": "systemd"}, "user@0.service": {"name": "user@0.service", "state": "running", "status": "active", "source": "systemd"}, "wpa_supplicant.service": {"name": "wpa_supplicant.service", "state": "running", "status": "enabled", "source": "systemd"}, "ypbind.service": {"name": "ypbind.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "yppasswdd.service": {"name": "yppasswdd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ypserv.service": {"name": "ypserv.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ypxfrd.service": {"name": "ypxfrd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "autovt@.service": {"name": "autovt@.service", "state": "unknown", "status": "alias", "source": "systemd"}, "chrony-wait.service": {"name": "chrony-wait.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "chronyd-restricted.service": {"name": "chronyd-restricted.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "cloud-init-hotplugd.service": {"name": "cloud-init-hotplugd.service", "state": "inactive", "status": "static", "source": "systemd"}, "console-getty.service": {"name": "console-getty.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "container-getty@.service": {"name": "container-getty@.service", "state": "unknown", "status": "static", "source": "systemd"}, "cpupower.service": {"name": "cpupower.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dbus-org.freedesktop.hostname1.service": {"name": "dbus-org.freedesktop.hostname1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.locale1.service": {"name": "dbus-org.freedesktop.locale1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.login1.service": {"name": "dbus-org.freedesktop.login1.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.nm-dispatcher.service": {"name": "dbus-org.freedesktop.nm-dispatcher.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.timedate1.service": {"name": "dbus-org.freedesktop.timedate1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus.service": {"name": "dbus.service", "state": "active", "status": "alias", "source": "systemd"}, "debug-shell.service": {"name": "debug-shell.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnf-system-upgrade-cleanup.service": {"name": "dnf-system-upgrade-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "dnf-system-upgrade.service": {"name": "dnf-system-upgrade.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnsmasq.service": {"name": "dnsmasq.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "firewalld.service": {"name": "firewalld.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fstrim.service": {"name": "fstrim.service", "state": "inactive", "status": "static", "source": "systemd"}, "getty@.service": {"name": "getty@.service", "state": "unknown", "status": "enabled", "source": "systemd"}, "grub-boot-indeterminate.service": {"name": "grub-boot-indeterminate.service", "state": "inactive", "status": "static", "source": "systemd"}, "grub2-systemd-integration.service": {"name": "grub2-systemd-integration.service", "state": "inactive", "status": "static", "source": "systemd"}, "hostapd.service": {"name": "hostapd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "kvm_stat.service": {"name": "kvm_stat.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "man-db-cache-update.service": {"name": "man-db-cache-update.service", "state": "inactive", "status": "static", "source": "systemd"}, "man-db-restart-cache-update.service": {"name": "man-db-restart-cache-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "modprobe@.service": {"name": "modprobe@.service", "state": "unknown", "status": "static", "source": "systemd"}, "nfs-blkmap.service": {"name": "nfs-blkmap.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nftables.service": {"name": "nftables.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nm-priv-helper.service": {"name": "nm-priv-helper.service", "state": "inactive", "status": "static", "source": "systemd"}, "oddjobd.service": {"name": "oddjobd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "pam_namespace.service": {"name": "pam_namespace.service", "state": "inactive", "status": "static", "source": "systemd"}, "qemu-guest-agent.service": {"name": "qemu-guest-agent.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "quotaon.service": {"name": "quotaon.service", "state": "inactive", "status": "static", "source": "systemd"}, "rdisc.service": {"name": "rdisc.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "rpmdb-rebuild.service": {"name": "rpmdb-rebuild.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "selinux-autorelabel.service": {"name": "selinux-autorelabel.service", "state": "inactive", "status": "static", "source": "systemd"}, "selinux-check-proper-disable.service": {"name": "selinux-check-proper-disable.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "serial-getty@.service": {"name": "serial-getty@.service", "state": "unknown", "status": "indirect", "source": "systemd"}, "sshd-keygen@.service": {"name": "sshd-keygen@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "sshd@.service": {"name": "sshd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "sssd-autofs.service": {"name": "sssd-autofs.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-nss.service": {"name": "sssd-nss.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pac.service": {"name": "sssd-pac.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pam.service": {"name": "sssd-pam.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-ssh.service": {"name": "sssd-ssh.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-sudo.service": {"name": "sssd-sudo.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "system-update-cleanup.service": {"name": "system-update-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-backlight@.service": {"name": "systemd-backlight@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-bless-boot.service": {"name": "systemd-bless-boot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-boot-check-no-failures.service": {"name": "systemd-boot-check-no-failures.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-coredump@.service": {"name": "systemd-coredump@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-exit.service": {"name": "systemd-exit.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-fsck@.service": {"name": "systemd-fsck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-growfs-root.service": {"name": "systemd-growfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-growfs@.service": {"name": "systemd-growfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-halt.service": {"name": "systemd-halt.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hibernate-resume@.service": {"name": "systemd-hibernate-resume@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-hibernate.service": {"name": "systemd-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hostnamed.service": {"name": "systemd-hostnamed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hybrid-sleep.service": {"name": "systemd-hybrid-sleep.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-journald@.service": {"name": "systemd-journald@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-kexec.service": {"name": "systemd-kexec.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-localed.service": {"name": "systemd-localed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs-root.service": {"name": "systemd-pcrfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs@.service": {"name": "systemd-pcrfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-poweroff.service": {"name": "systemd-poweroff.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pstore.service": {"name": "systemd-pstore.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-quotacheck.service": {"name": "systemd-quotacheck.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-reboot.service": {"name": "systemd-reboot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend-then-hibernate.service": {"name": "systemd-suspend-then-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend.service": {"name": "systemd-suspend.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-sysupdate-reboot.service": {"name": "systemd-sysupdate-reboot.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-sysupdate.service": {"name": "systemd-sysupdate.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-timedated.service": {"name": "systemd-timedated.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-volatile-root.service": {"name": "systemd-volatile-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "teamd@.service": {"name": "teamd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user-runtime-dir@.service": {"name": "user-runtime-dir@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user@.service": {"name": "user@.service", "state": "unknown", "status": "static", "source": "systemd"}}}, "invocation": {"module_args": {}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204091.10287: done with _execute_module (service_facts, {'_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'service_facts', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204089.658723-15364-162876841556701/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204091.10304: _low_level_execute_command(): starting 13118 1727204091.10359: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204089.658723-15364-162876841556701/ > /dev/null 2>&1 && sleep 0' 13118 1727204091.11160: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204091.11180: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.11196: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204091.11219: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.11267: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204091.11281: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204091.11297: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.11316: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204091.11333: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204091.11349: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204091.11363: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.11381: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204091.11398: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.11410: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204091.11422: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204091.11438: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.11520: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204091.11546: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204091.11574: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204091.11648: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204091.13768: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204091.13979: stderr chunk (state=3): >>><<< 13118 1727204091.13991: stdout chunk (state=3): >>><<< 13118 1727204091.14171: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204091.14175: handler run complete 13118 1727204091.14291: variable 'ansible_facts' from source: unknown 13118 1727204091.14384: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204091.14897: variable 'ansible_facts' from source: unknown 13118 1727204091.15127: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204091.15340: attempt loop complete, returning result 13118 1727204091.15350: _execute() done 13118 1727204091.15357: dumping result to json 13118 1727204091.15425: done dumping result, returning 13118 1727204091.15442: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check which services are running [0affcd87-79f5-56a3-0a64-000000000496] 13118 1727204091.15452: sending task result for task 0affcd87-79f5-56a3-0a64-000000000496 ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13118 1727204091.16760: no more pending results, returning what we have 13118 1727204091.16766: results queue empty 13118 1727204091.16767: checking for any_errors_fatal 13118 1727204091.16785: done checking for any_errors_fatal 13118 1727204091.16786: checking for max_fail_percentage 13118 1727204091.16788: done checking for max_fail_percentage 13118 1727204091.16789: checking to see if all hosts have failed and the running result is not ok 13118 1727204091.16790: done checking to see if all hosts have failed 13118 1727204091.16791: getting the remaining hosts for this loop 13118 1727204091.16792: done getting the remaining hosts for this loop 13118 1727204091.16796: getting the next task for host managed-node2 13118 1727204091.16804: done getting next task for host managed-node2 13118 1727204091.16810: ^ task is: TASK: fedora.linux_system_roles.network : Check which packages are installed 13118 1727204091.16816: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204091.16829: getting variables 13118 1727204091.16834: in VariableManager get_vars() 13118 1727204091.16875: Calling all_inventory to load vars for managed-node2 13118 1727204091.16878: Calling groups_inventory to load vars for managed-node2 13118 1727204091.16881: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204091.16900: Calling all_plugins_play to load vars for managed-node2 13118 1727204091.16903: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204091.16906: Calling groups_plugins_play to load vars for managed-node2 13118 1727204091.17984: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000496 13118 1727204091.17988: WORKER PROCESS EXITING 13118 1727204091.19997: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204091.22945: done with get_vars() 13118 1727204091.22986: done getting variables TASK [fedora.linux_system_roles.network : Check which packages are installed] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:26 Tuesday 24 September 2024 14:54:51 -0400 (0:00:01.635) 0:00:28.530 ***** 13118 1727204091.23100: entering _queue_task() for managed-node2/package_facts 13118 1727204091.23660: worker is 1 (out of 1 available) 13118 1727204091.23676: exiting _queue_task() for managed-node2/package_facts 13118 1727204091.23689: done queuing things up, now waiting for results queue to drain 13118 1727204091.23690: waiting for pending results... 13118 1727204091.24449: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check which packages are installed 13118 1727204091.24640: in run() - task 0affcd87-79f5-56a3-0a64-000000000497 13118 1727204091.24661: variable 'ansible_search_path' from source: unknown 13118 1727204091.24673: variable 'ansible_search_path' from source: unknown 13118 1727204091.24710: calling self._execute() 13118 1727204091.24829: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204091.24844: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204091.24859: variable 'omit' from source: magic vars 13118 1727204091.25260: variable 'ansible_distribution_major_version' from source: facts 13118 1727204091.25285: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204091.25299: variable 'omit' from source: magic vars 13118 1727204091.25393: variable 'omit' from source: magic vars 13118 1727204091.25444: variable 'omit' from source: magic vars 13118 1727204091.25492: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204091.25542: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204091.25569: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204091.25591: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204091.25605: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204091.25649: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204091.25657: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204091.25665: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204091.25778: Set connection var ansible_timeout to 10 13118 1727204091.25795: Set connection var ansible_pipelining to False 13118 1727204091.25801: Set connection var ansible_connection to ssh 13118 1727204091.25810: Set connection var ansible_shell_executable to /bin/sh 13118 1727204091.25818: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204091.25824: Set connection var ansible_shell_type to sh 13118 1727204091.25859: variable 'ansible_shell_executable' from source: unknown 13118 1727204091.25871: variable 'ansible_connection' from source: unknown 13118 1727204091.25878: variable 'ansible_module_compression' from source: unknown 13118 1727204091.25884: variable 'ansible_shell_type' from source: unknown 13118 1727204091.25890: variable 'ansible_shell_executable' from source: unknown 13118 1727204091.25895: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204091.25903: variable 'ansible_pipelining' from source: unknown 13118 1727204091.25909: variable 'ansible_timeout' from source: unknown 13118 1727204091.25916: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204091.26236: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204091.26252: variable 'omit' from source: magic vars 13118 1727204091.26260: starting attempt loop 13118 1727204091.26272: running the handler 13118 1727204091.26291: _low_level_execute_command(): starting 13118 1727204091.26306: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204091.27155: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204091.27182: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.27212: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204091.27244: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.27462: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204091.27477: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204091.27498: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.27517: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204091.27529: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204091.27544: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204091.27557: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.27574: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204091.27592: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.27615: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204091.27627: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204091.27644: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.27729: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204091.27750: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204091.27769: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204091.27880: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204091.29502: stdout chunk (state=3): >>>/root <<< 13118 1727204091.29704: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204091.29708: stdout chunk (state=3): >>><<< 13118 1727204091.29711: stderr chunk (state=3): >>><<< 13118 1727204091.29836: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204091.29840: _low_level_execute_command(): starting 13118 1727204091.29843: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204091.2973328-15447-256388152982611 `" && echo ansible-tmp-1727204091.2973328-15447-256388152982611="` echo /root/.ansible/tmp/ansible-tmp-1727204091.2973328-15447-256388152982611 `" ) && sleep 0' 13118 1727204091.30494: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204091.30512: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.30528: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204091.30547: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.30594: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204091.30613: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204091.30626: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.30643: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204091.30654: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204091.30671: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204091.30684: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.30697: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204091.30715: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.30729: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204091.30740: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204091.30752: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.30837: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204091.30854: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204091.30868: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204091.31054: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204091.32853: stdout chunk (state=3): >>>ansible-tmp-1727204091.2973328-15447-256388152982611=/root/.ansible/tmp/ansible-tmp-1727204091.2973328-15447-256388152982611 <<< 13118 1727204091.33071: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204091.33076: stdout chunk (state=3): >>><<< 13118 1727204091.33094: stderr chunk (state=3): >>><<< 13118 1727204091.33472: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204091.2973328-15447-256388152982611=/root/.ansible/tmp/ansible-tmp-1727204091.2973328-15447-256388152982611 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204091.33476: variable 'ansible_module_compression' from source: unknown 13118 1727204091.33478: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.package_facts-ZIP_DEFLATED 13118 1727204091.33480: variable 'ansible_facts' from source: unknown 13118 1727204091.33482: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204091.2973328-15447-256388152982611/AnsiballZ_package_facts.py 13118 1727204091.33621: Sending initial data 13118 1727204091.33624: Sent initial data (162 bytes) 13118 1727204091.34606: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204091.34622: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.34637: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204091.34655: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.34707: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204091.34723: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204091.34738: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.34755: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204091.34769: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204091.34787: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204091.34802: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.34816: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204091.34830: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.34842: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204091.34852: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204091.34866: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.34942: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204091.34968: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204091.34986: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204091.35062: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204091.36791: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204091.36826: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204091.36869: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpddgjxvlf /root/.ansible/tmp/ansible-tmp-1727204091.2973328-15447-256388152982611/AnsiballZ_package_facts.py <<< 13118 1727204091.36893: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204091.39144: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204091.39449: stderr chunk (state=3): >>><<< 13118 1727204091.39454: stdout chunk (state=3): >>><<< 13118 1727204091.39456: done transferring module to remote 13118 1727204091.39462: _low_level_execute_command(): starting 13118 1727204091.39480: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204091.2973328-15447-256388152982611/ /root/.ansible/tmp/ansible-tmp-1727204091.2973328-15447-256388152982611/AnsiballZ_package_facts.py && sleep 0' 13118 1727204091.40082: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204091.40099: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.40114: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204091.40136: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.40183: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204091.40195: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204091.40207: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.40224: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204091.40235: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204091.40249: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204091.40261: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.40276: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204091.40290: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.40301: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204091.40311: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204091.40323: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.40399: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204091.40422: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204091.40438: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204091.40505: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204091.42284: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204091.42402: stderr chunk (state=3): >>><<< 13118 1727204091.42422: stdout chunk (state=3): >>><<< 13118 1727204091.42470: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204091.42474: _low_level_execute_command(): starting 13118 1727204091.42476: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204091.2973328-15447-256388152982611/AnsiballZ_package_facts.py && sleep 0' 13118 1727204091.43126: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204091.43141: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.43155: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204091.43177: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.43231: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204091.43245: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204091.43260: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.43280: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204091.43293: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204091.43309: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204091.43326: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.43341: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204091.43357: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.43374: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204091.43386: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204091.43400: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.43492: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204091.43517: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204091.43543: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204091.43626: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204091.90016: stdout chunk (state=3): >>> {"ansible_facts": {"packages": {"libgcc": [{"name": "libgcc", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "linux-firmware-whence": [{"name": "linux-firmware-whence", "version": "20240905", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "tzdata": [{"name": "tzdata", "version": "2024a", "release": "2.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "linux-firmware": [{"name": "linux-firmware", "version": "20240905", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "gawk-all-langpacks": [{"name": "gawk-all-langpacks", "version": "5.1.0", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools-wheel": [{"name": "python3-setuptools-wheel", "version": "53.0.0", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "publicsuffix-list-dafsa": [{"name": "publicsuffix-list-dafsa", "version": "20210518", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "pcre2-syntax": [{"name": "pcre2-syntax", "version": "10.40", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "ncurses-base": [{"name": "ncurses-base", "version": "6.2", "release": "10.20210508.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libssh-config": [{"name": "libssh-config", "version": "0.10.4", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libreport-filesystem": [{"name": "libreport-filesystem", "version": "2.15.2", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf-data": [{"name": "dnf-data", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-misc": [{"name": "kbd-misc", "version": "2.4.0", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-legacy": [{"name": "kbd-legacy", "version": "2.4.0", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "hwdata": [{"name": "hwdata", "version": "0.348", "release": "9.15.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "fonts-filesystem": [{"name": "fonts-filesystem", "version": "2.0.5", "release": "7.el9.1", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dejavu-sans-fonts": [{"name": "dejavu-sans-fonts", "version": "2.37", "release": "18.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-core-font-en": [{"name": "langpacks-core-font-en", "version": "3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "firewalld-filesystem": [{"name": "firewalld-filesystem", "version": "1.3.4", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "coreutils-common": [{"name": "coreutils-common", "version": "8.32", "release": "36.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "centos-gpg-keys": [{"name": "centos-gpg-keys", "version": "9.0", "release": "26.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-repos": [{"name": "centos-stream-repos", "version": "9.0", "release": "26.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-release": [{"name": "centos-stream-release", "version": "9.0", "release": "26.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "setup": [{"name": "setup", "version": "2.13.7", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "filesystem": [{"name": "filesystem", "version": "3.16", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "basesystem": [{"name": "basesystem", "version": "11", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "glibc-gconv-extra": [{"name": "glibc-gconv-extra", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-langpack-en": [{"name": "glibc-langpack-en", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-common": [{"name": "glibc-common", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gli<<< 13118 1727204091.90063: stdout chunk (state=3): >>>bc": [{"name": "glibc", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses-libs": [{"name": "ncurses-libs", "version": "6.2", "release": "10.20210508.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bash": [{"name": "bash", "version": "5.1.8", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib": [{"name": "zlib", "version": "1.2.11", "release": "41.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-libs": [{"name": "xz-libs", "version": "5.2.5", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd": [{"name": "libzstd", "version": "1.5.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap": [{"name": "libcap", "version": "2.48", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "popt": [{"name": "popt", "version": "1.18", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bzip2-libs": [{"name": "bzip2-libs", "version": "1.0.8", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt": [{"name": "libxcrypt", "version": "4.4.18", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libuuid": [{"name": "libuuid", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sqlite-libs": [{"name": "sqlite-libs", "version": "3.34.1", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcom_err": [{"name": "libcom_err", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libstdc++": [{"name": "libstdc++", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmnl": [{"name": "libmnl", "version": "1.0.4", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf": [{"name": "elfutils-libelf", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxml2": [{"name": "libxml2", "version": "2.9.13", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies": [{"name": "crypto-policies", "version": "20240828", "release": "2.git626aa59.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "readline": [{"name": "readline", "version": "8.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "alternatives": [{"name": "alternatives", "version": "1.24", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng": [{"name": "libcap-ng", "version": "0.8.2", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit-libs": [{"name": "audit-libs", "version": "3.1.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libunistring": [{"name": "libunistring", "version": "0.9.10", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lua-libs": [{"name": "lua-libs", "version": "5.4.4", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libffi": [{"name": "libffi", "version": "3.4.2", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgpg-error": [{"name": "libgpg-error", "version": "1.42", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtalloc": [{"name": "libtalloc", "version": "2.4.2", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libidn2": [{"name": "libidn2", "version": "2.3.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gmp": [{"name": "gmp", "version": "6.2.0", "release": "13.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "keyutils-libs": [{"name": "keyutils-libs", "version": "1.6.3", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libattr": [{"name": "libattr", "version": "2.5.1", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libacl": [{"name": "libacl", "version": "2.3.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3": [{"name": "libnl3", "version": "3.9.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsepol": [{"name": "libsepol", "version": "3.6", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsmartcols": [{"name": "libsmartcols", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lz4-libs": [{"name": "lz4-libs", "version": "1.9.3", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre2": [{"name": "pcre2", "version": "10.40", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux": [{"name": "libselinux", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sed": [{"name": "sed", "version": "4.8", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "findutils": [{"name": "findutils", "version": "4.8.0", "release": "7.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libsemanage": [{"name": "libsemanage", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "shadow-utils": [{"name": "shadow-utils", "version": "4.9", "release": "9.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libtevent": [{"name": "libtevent", "version": "0.16.1", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgcrypt": [{"name": "libgcrypt", "version": "1.10.0", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file-libs": [{"name": "file-libs", "version": "5.39", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libedit": [{"name": "libedit", "version": "3.1", "release": "38.20210216cvs.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "psmisc": [{"name": "psmisc", "version": "23.4", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "expat": [{"name": "expat", "version": "2.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm-libs": [{"name": "gdbm-libs", "version": "1.23", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "jansson": [{"name": "jansson", "version": "2.14", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "json-c": [{"name": "json-c", "version": "0.14", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtasn1": [{"name": "libtasn1", "version": "4.16.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit": [{"name": "p11-kit", "version": "0.25.3", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtdb": [{"name": "libtdb", "version": "1.4.10", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit-trust": [{"name": "p11-kit-trust", "version": "0.25.3", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file": [{"name": "file", "version": "5.39", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libassuan": [{"name": "libassuan", "version": "2.5.5", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbpf": [{"name": "libbpf", "version": "1.4.0", "release": "1.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libnftnl": [{"name": "libnftnl", "version": "1.2.6", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse-libs": [{"name": "fuse-libs", "version": "2.9.9", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbasicobjects": [{"name": "libbasicobjects", "version": "0.1.1", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcollection": [{"name": "libcollection", "version": "0.7.0", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdb": [{"name": "libdb", "version": "5.3.28", "release": "54.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute": [{"name": "iproute", "version": "6.2.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdhash": [{"name": "libdhash", "version": "0.5.0", "release": <<< 13118 1727204091.90151: stdout chunk (state=3): >>>"53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgomp": [{"name": "libgomp", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libref_array": [{"name": "libref_array", "version": "0.1.5", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libseccomp": [{"name": "libseccomp", "version": "2.5.2", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsigsegv": [{"name": "libsigsegv", "version": "2.13", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_idmap": [{"name": "libsss_idmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lzo": [{"name": "lzo", "version": "2.10", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "numactl-libs": [{"name": "numactl-libs", "version": "2.0.18", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre": [{"name": "pcre", "version": "8.44", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grep": [{"name": "grep", "version": "3.6", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssl-libs": [{"name": "openssl-libs", "version": "3.2.2", "release": "6.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "coreutils": [{"name": "coreutils", "version": "8.32", "release": "36.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ca-certificates": [{"name": "ca-certificates", "version": "2024.2.69_v8.0.303", "release": "91.4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "systemd-libs": [{"name": "systemd-libs", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libblkid": [{"name": "libblkid", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-libs": [{"name": "dbus-libs", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libmount": [{"name": "libmount", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux-core": [{"name": "util-linux-core", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfdisk": [{"name": "libfdisk", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gzip": [{"name": "gzip", "version": "1.12", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod": [{"name": "kmod", "version": "28", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod-libs": [{"name": "kmod-libs", "version": "28", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib": [{"name": "cracklib", "version": "2.9.6", "release": "27.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "which": [{"name": "which", "version": "2.21", "release": "29.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib-dicts": [{"name": "cracklib-dicts", "version": "2.9.6", "release": "27.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-tools": [{"name": "dbus-tools", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "procps-ng": [{"name": "procps-ng", "version": "3.3.17", "release": "14.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssl": [{"name": "openssl", "version": "3.2.2", "release": "6.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libarchive": [{"name": "libarchive", "version": "3.5.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libevent": [{"name": "libevent", "version": "2.1.12", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_certmap": [{"name": "libsss_certmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz": [{"name": "xz", "version": "5.2.5", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "squashfs-tools": [{"name": "squashfs-tools", "version": "4.4", "release": "10.git1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcomps": [{"name": "libcomps", "version": "0.1.18", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libutempter": [{"name": "libutempter", "version": "1.2.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux-utils": [{"name": "libselinux-utils", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3-cli": [{"name": "libnl3-cli", "version": "3.9.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libteam": [{"name": "libteam", "version": "1.31", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "acl": [{"name": "acl", "version": "2.3.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-libs": [{"name": "gettext-libs", "version": "0.21", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext": [{"name": "gettext", "version": "0.21", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "attr": [{"name": "attr", "version": "2.5.1", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils": [{"name": "keyutils", "version": "1.6.3", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpfr": [{"name": "mpfr", "version": "4.1.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gawk": [{"name": "gawk", "version": "5.1.0", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpsl": [{"name": "libpsl", "version": "0.21.1", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libksba": [{"name": "libksba", "version": "1.5.1", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ethtool": [{"name": "ethtool", "version": "6.2", "release": "1.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "ipset-libs": [{"name": "ipset-libs", "version": "7.11", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ipset": [{"name": "ipset", "version": "7.11", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "groff-base": [{"name": "groff-base", "version": "1.22.4", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "snappy": [{"name": "snappy", "version": "1.1.8", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs-libs": [{"name": "e2fsprogs-libs", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libss": [{"name": "libss", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt-compat": [{"name": "libxcrypt-compat", "version": "4.4.18", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip-wheel": [{"name": "python3-pip-wheel", "version": "21.3.1", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python-unversioned-command": [{"name": "python-unversioned-command", "version": "3.9.19", "release": "8.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3": [{"name": "python3", "version": "3.9.19", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libs": [{"name": "python3-libs", "version": "3.9.19", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libcomps": [{"name": "python3-libcomps", "version": "0.1.18", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-six": [{"name": "python3-six", "version": "1.15.0", "release": "9.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dateutil": [{"name": "python3-dateutil", "version": "2.8.1", "release": "7.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "python3-systemd": [{"name": "python3-systemd", "version": "234", "release": "19.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng-python3": [{"name": "libcap-ng-python3", "version": "0.8.2", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pigz": [{"name": "pigz", "version": "2.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "hostname": [{"name": "hostname", "version": "3.23", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools-libs": [{"name": "kernel-tools-libs", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "less": [{"name": "less", "version": "590", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-rpm-macros": [{"name": "systemd-rpm-macros", "version": "252", "release": "47.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "c-ares": [{"name": "c-ares", "version": "1.19.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpio": [{"name": "cpio", "version": "2.13", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "diffutils": [{"name": "diffutils", "version": "3.7", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "inih": [{"name": "inih", "version": "49", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbrotli": [{"name": "libbrotli", "version": "1.0.9", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcbor": [{"name": "libcbor", "version": "0.7.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdaemon": [{"name": "libdaemon", "version": "0.14", "release": "23.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "teamd": [{"name": "teamd", "version": "1.31", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libeconf": [{"name": "libeconf", "version": "0.4.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpwquality": [{"name": "libpwquality", "version": "1.4.4", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pam": [{"name": "pam", "version": "1.5.1", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux": [{"name": "util-linux", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus": [{"name": "dbus", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "systemd-pam": [{"name": "systemd-pam", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd": [{"name": "systemd", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-common": [{"name": "dbus-common", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dbus-broker": [{"name": "dbus-broker", "version": "28", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-common": [{"name": "grub2-common", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "cronie-anacron": [{"name": "cronie-anacron", "version": "1.5.7", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cronie": [{"name": "cronie", "version": "1.5.7", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crontabs": [{"name": "crontabs", "version": "1.11", "release": "26.20190603git.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "openssh": [{"name": "openssh", "version": "8.7p1", "release": "43.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc-modules": [{"name": "grub2-pc-modules", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "authselect-libs": [{"name": "authselect-libs", "version": "1.2.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-libs": [{"name": "device-mapper-libs", "version": "1.02.198", "release": "2.el9", "epoch": 9, "arch": "x86_64", "source": "rpm"}], "device-mapper": [{"name": "device-mapper", "version": "1.02.198", "release": "2.el9", "epoch": 9, "arch": "x86_64", "source": "rpm"}], "grub2-tools-minimal": [{"name": "grub2-tools-minimal", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "cryptsetup-libs": [{"name": "cryptsetup-libs", "version": "2.7.2", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kpartx": [{"name": "kpartx", "version": "0.8.7", "release": "32.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-default-yama-scope": [{"name": "elfutils-default-yama-scope", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "elfutils-libs": [{"name": "elfutils-libs", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-service": [{"name": "initscripts-service", "version": "10.11.8", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iputils": [{"name": "iputils", "version": "20210202", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi": [{"name": "libkcapi", "version": "1.4.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hmaccalc": [{"name": "libkcapi-hmaccalc", "version": "1.4.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "logrotate": [{"name": "logrotate", "version": "3.18.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oddjob": [{"name": "oddjob", "version": "0.34.7", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oddjob-mkhomedir": [{"name": "oddjob-mkhomedir", "version": "0.34.7", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect": [{"name": "authselect", "version": "1.2.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kbd": [{"name": "kbd", "version": "2.4.0", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libndp": [{"name": "libndp", "version": "1.9", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfnetlink": [{"name": "libnfnetlink", "version": "1.0.1", "release": "21.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnetfilter_conntrack": [{"name": "libnetfilter_conntrack", "version": "1.0.9", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-libs": [{"name": "iptables-libs", "version": "1.8.10", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-nft": [{"name": "iptables-nft", "version": "1.8.10", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nftables": [{"name": "nftables", "version": "1.0.9", "release": "3.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-nftables": [{"name": "python3-nftables", "version": "1.0.9", "release": "3.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libnghttp2": [{"name": "libnghttp2", "version": "1.43.0", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpath_utils": [{"name": "libpath_utils", "version": "0.2.1", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libini_config": [{"name": "libini_config", "version": "1.3.1", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpipeline": [{"name": "libpipeline", "version": "1.5.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_nss_idmap": [{"name": "libsss_nss_idmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_sudo": [{"name": "libsss_sudo", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto": [{"name": "libverto", "version": "0.3.2", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "krb5-libs": [{"name": "krb5-libs", "version": "1.21.1", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-lib": [{"name": "cyrus-sasl-lib", "version": "2.1.27", "release": "21.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openldap": [{"name": "openldap", "version": "2.6.6", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libssh": [{"name": "libssh", "version": "0.10.4", "release": "13.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcurl": [{"name": "libcurl", "version": "7.76.1", "release": "31.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-debuginfod-client": [{"name": "elfutils-debuginfod-client", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils-gold": [{"name": "binutils-gold", "version": "2.35.2", "release": "54.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils": [{"name": "binutils", "version": "2.35.2", "release": "54.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tss": [{"name": "tpm2-tss", "version": "3.2.3", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-udev": [{"name": "systemd-udev", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut": [{"name": "dracut", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-modules-core": [{"name": "kernel-modules-core", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-core": [{"name": "kernel-core", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-modules": [{"name": "kernel-modules", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-squash": [{"name": "dracut-squash", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfido2": [{"name": "libfido2", "version": "1.13.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "os-prober": [{"name": "os-prober", "version": "1.77", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools": [{"name": "grub2-tools", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "ima-evm-utils": [{"name": "ima-evm-utils", "version": "1.5", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "curl": [{"name": "curl", "version": "7.76.1", "release": "31.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm": [{"name": "rpm", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-libs": [{"name": "rpm-libs", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grubby": [{"name": "grubby", "version": "8.40", "release": "63.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsolv": [{"name": "libsolv", "version": "0.7.24", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "policycoreutils": [{"name": "policycoreutils", "version": "3.6", "release": "2.1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "selinux-policy": [{"name": "selinux-policy", "version": "38.1.45", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "selinux-policy-targeted": [{"name": "selinux-policy-targeted", "version": "38.1.45", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpm-build-libs": [{"name": "rpm-build-libs", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-systemd-inhibit": [{"name": "rpm-plugin-systemd-inhibit", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-client": [{"name": "sssd-client", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libyaml": [{"name": "libyaml", "version": "0.2.5", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lmdb-libs": [{"name": "lmdb-libs", "version": "0.9.29", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libldb": [{"name": "libldb", "version": "2.9.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-common": [{"name": "sssd-common", "version": "2.9.5", "rel<<< 13118 1727204091.90200: stdout chunk (state=3): >>>ease": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nettle": [{"name": "nettle", "version": "3.9.1", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnutls": [{"name": "gnutls", "version": "3.8.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glib2": [{"name": "glib2", "version": "2.68.4", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dbus": [{"name": "python3-dbus", "version": "1.2.18", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-libnm": [{"name": "NetworkManager-libnm", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "NetworkManager": [{"name": "NetworkManager", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libmodulemd": [{"name": "libmodulemd", "version": "2.13.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gobject-introspection": [{"name": "gobject-introspection", "version": "1.68.0", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gobject-base-noarch": [{"name": "python3-gobject-base-noarch", "version": "3.40.1", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-gobject-base": [{"name": "python3-gobject-base", "version": "3.40.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-firewall": [{"name": "python3-firewall", "version": "1.3.4", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libuser": [{"name": "libuser", "version": "0.63", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "npth": [{"name": "npth", "version": "1.6", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2": [{"name": "gnupg2", "version": "2.3.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gpgme": [{"name": "gpgme", "version": "1.15.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "librepo": [{"name": "librepo", "version": "1.14.5", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdnf": [{"name": "libdnf", "version": "0.69.0", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libdnf": [{"name": "python3-libdnf", "version": "0.69.0", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-hawkey": [{"name": "python3-hawkey", "version": "0.69.0", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gpg": [{"name": "python3-gpg", "version": "1.15.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sign-libs": [{"name": "rpm-sign-libs", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpm": [{"name": "python3-rpm", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dnf": [{"name": "python3-dnf", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf": [{"name": "dnf", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dnf-plugins-core": [{"name": "python3-dnf-plugins-core", "version": "4.3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "oniguruma": [{"name": "oniguruma", "version": "6.9.6", "release": "1.el9.6", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jq": [{"name": "jq", "version": "1.6", "release": "17.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-network": [{"name": "dracut-network", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pciutils-libs": [{"name": "pciutils-libs", "version": "3.7.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sg3_utils-libs": [{"name": "sg3_utils-libs", "version": "1.47", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "slang": [{"name": "slang", "version": "2.3.2", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "newt": [{"name": "newt", "version": "0.52.21", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "userspace-rcu": [{"name": "userspace-rcu", "version": "0.12.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libestr": [{"name": "libestr", "version": "0.1.11", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfastjson": [{"name": "libfastjson", "version": "0.99.9", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsyslog-logrotate": [{"name": "rsyslog-logrotate", "version": "8.2310.0", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsyslog": [{"name": "rsyslog", "version": "8.2310.0", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "liburing": [{"name": "liburing", "version": "2.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "langpacks-core-en": [{"name": "langpacks-core-en", "version": "3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-en": [{"name": "langpacks-en", "version": "3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "qemu-guest-agent": [{"name": "qemu-guest-agent", "version": "9.0.0", "release": "10.el9", "epoch": 17, "arch": "x86_64", "source": "rpm"}], "xfsprogs": [{"name": "xfsprogs", "version": "6.4.0", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-tui": [{"name": "NetworkManager-tui", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "sg3_utils": [{"name": "sg3_utils", "version": "1.47", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools": [{"name": "kernel-tools", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kexec-tools": [{"name": "kexec-tools", "version": "2.0.27", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dnf-plugins-core": [{"name": "dnf-plugins-core", "version": "4.3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "yum": [{"name": "yum", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "passwd": [{"name": "passwd", "version": "0.80", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "firewalld": [{"name": "firewalld", "version": "1.3.4", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "NetworkManager-team": [{"name": "NetworkManager-team", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "initscripts-rename-device": [{"name": "initscripts-rename-device", "version": "10.11.8", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "irqbalance": [{"name": "irqbalance", "version": "1.9.4", "release": "1.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "chrony": [{"name": "chrony", "version": "4.6", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-kcm": [{"name": "sssd-kcm", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-selinux": [{"name": "rpm-plugin-selinux", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies-scripts": [{"name": "crypto-policies-scripts", "version": "20240828", "release": "2.git626aa59.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpm-plugin-audit": [{"name": "rpm-plugin-audit", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc": [{"name": "grub2-pc", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "openssh-clients": [{"name": "openssh-clients", "version": "8.7p1", "release": "43.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel": [{"name": "kernel", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-config-rescue": [{"name": "dracut-config-rescue", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo": [{"name": "sudo", "version": "1.9.5p2", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit": [{"name": "audit", "version": "3.1.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-server": [{"name": "openssh-server", "version": "8.7p1", "release": "43.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "man-db": [{"name": "man-db", "version": "2.9.3", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute-tc": [{"name": "iproute-tc", "version": "6.2.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect-compat": [{"name": "authselect-compat", "version": "1.2.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "parted": [{"name": "parted", "version": "3.5", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "microcode_ctl": [{"name": "microcode_ctl", "version": "20240531", "release": "1.el9", "epoch": 4, "arch": "noarch", "source": "rpm"}], "python3-libselinux": [{"name": "python3-libselinux", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs": [{"name": "e2fsprogs", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "prefixdevname": [{"name": "prefixdevname", "version": "0.1.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-minimal": [{"name": "vim-minimal", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "lshw": [{"name": "lshw", "version": "B.02.19.2", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses": [{"name": "ncurses", "version": "6.2", "release": "10.20210508.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsysfs": [{"name": "libsysfs", "version": "2.1.1", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lsscsi": [{"name": "lsscsi", "version": "0.32", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iwl100-firmware": [{"name": "iwl100-firmware", "version": "39.31.5.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl1000-firmware": [{"name": "iwl1000-firmware", "version": "39.31.5.1", "release": "146.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "iwl105-firmware": [{"name": "iwl105-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl135-firmware": [{"name": "iwl135-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl2000-firmware": [{"name": "iwl2000-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl2030-firmware": [{"name": "iwl2030-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl3160-firmware": [{"name": "iwl3160-firmware", "version": "25.30.13.0", "release": "146.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "iwl5000-firmware": [{"name": "iwl5000-firmware", "version": "8.83.5.1_1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl5150-firmware": [{"name": "iwl5150-firmware", "version": "8.24.2.2", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl6000g2a-firmware": [{"name": "iwl6000g2a-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl6050-firmware": [{"name": "iwl6050-firmware", "version": "41.28.5.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl7260-firmware": [{"name": "iwl7260-firmware", "version": "25.30.13.0", "release": "146.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "rootfiles"<<< 13118 1727204091.90227: stdout chunk (state=3): >>>: [{"name": "rootfiles", "version": "8.1", "release": "31.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "gpg-pubkey": [{"name": "gpg-pubkey", "version": "3228467c", "release": "613798eb", "epoch": null, "arch": null, "source": "rpm"}, {"name": "gpg-pubkey", "version": "8483c65d", "release": "5ccc5b19", "epoch": null, "arch": null, "source": "rpm"}], "epel-release": [{"name": "epel-release", "version": "9", "release": "8.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "nspr": [{"name": "nspr", "version": "4.35.0", "release": "14.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "boost-system": [{"name": "boost-system", "version": "1.75.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-util": [{"name": "nss-util", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "make": [{"name": "make", "version": "4.3", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "m4": [{"name": "m4", "version": "1.4.19", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmpc": [{"name": "libmpc", "version": "1.2.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "unzip": [{"name": "unzip", "version": "6.0", "release": "57.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "avahi-libs": [{"name": "avahi-libs", "version": "0.8", "release": "21.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zip": [{"name": "zip", "version": "3.0", "release": "35.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpp": [{"name": "cpp", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bison": [{"name": "bison", "version": "3.7.4", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "flex": [{"name": "flex", "version": "2.6.4", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-softokn-freebl": [{"name": "nss-softokn-freebl", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-softokn": [{"name": "nss-softokn", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss": [{"name": "nss", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-sysinit": [{"name": "nss-sysinit", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "boost-filesystem": [{"name": "boost-filesystem", "version": "1.75.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "boost-thread": [{"name": "boost-thread", "version": "1.75.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Digest": [{"name": "perl-Digest", "version": "1.19", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Digest-MD5": [{"name": "perl-Digest-MD5", "version": "2.58", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-B": [{"name": "perl-B", "version": "1.80", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-FileHandle": [{"name": "perl-FileHandle", "version": "2.03", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Data-Dumper": [{"name": "perl-Data-Dumper", "version": "2.174", "release": "462.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-libnet": [{"name": "perl-libnet", "version": "3.13", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-base": [{"name": "perl-base", "version": "2.27", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-URI": [{"name": "perl-URI", "version": "5.09", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-AutoLoader": [{"name": "perl-AutoLoader", "version": "5.74", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Mozilla-CA": [{"name": "perl-Mozilla-CA", "version": "20200520", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-if": [{"name": "perl-if", "version": "0.60.800", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-IP": [{"name": "perl-IO-Socket-IP", "version": "0.41", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Time-Local": [{"name": "perl-Time-Local", "version": "1.300", "release": "7.el9", "epoch": 2, "arch": "noarch", "source": "rpm"}], "perl-File-Path": [{"name": "perl-File-Path", "version": "2.18", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Escapes": [{"name": "perl-Pod-Escapes", "version": "1.07", "release": "460.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Text-Tabs+Wrap": [{"name": "perl-Text-Tabs+Wrap", "version": "2013.0523", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-SSL": [{"name": "perl-IO-Socket-SSL", "version": "2.073", "release": "2.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Net-SSLeay": [{"name": "perl-Net-SSLeay", "version": "1.94", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Class-Struct": [{"name": "perl-Class-Struct", "version": "0.66", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-POSIX": [{"name": "perl-POSIX", "version": "1.94", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Term-ANSIColor": [{"name": "perl-Term-ANSIColor", "version": "5.01", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-IPC-Open3": [{"name": "perl-IPC-Open3", "version": "1.21", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-subs": [{"name": "perl-subs", "version": "1.03", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-Temp": [{"name": "perl-File-Temp", "version": "0.231.100", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Term-Cap": [{"name": "perl-Term-Cap", "version": "1.17", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Simple": [{"name": "perl-Pod-Simple", "version": "3.42", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-HTTP-Tiny": [{"name": "perl-HTTP-Tiny", "version": "0.076", "release": "462.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Socket": [{"name": "perl-Socket", "version": "2.031", "release": "4.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-SelectSaver": [{"name": "perl-SelectSaver", "version": "1.02", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Symbol": [{"name": "perl-Symbol", "version": "1.08", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-stat": [{"name": "perl-File-stat", "version": "1.09", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-podlators": [{"name": "perl-podlators", "version": "4.14", "release": "460.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Pod-Perldoc": [{"name": "perl-Pod-Perldoc", "version": "3.28.01", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Fcntl": [{"name": "perl-Fcntl", "version": "1.13", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Text-ParseWords": [{"name": "perl-Text-ParseWords", "version": "3.30", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-mro": [{"name": "perl-mro", "version": "1.23", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IO": [{"name": "perl-IO", "version": "1.43", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-overloading": [{"name": "perl-overloading", "version": "0.02", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Pod-Usage": [{"name": "perl-Pod-Usage", "version": "2.01", "release": "4.el9", "epoch": 4, "arch": "noarch", "source": "rpm"}], "perl-Errno": [{"name": "perl-Errno", "version": "1.30", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Basename": [{"name": "perl-File-Basename", "version": "2.85", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Std": [{"name": "perl-Getopt-Std", "version": "1.12", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-MIME-Base64": [{"name": "perl-MIME-Base64", "version": "3.16", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Scalar-List-Utils": [{"name": "perl-Scalar-List-Utils", "version": "1.56", "release": "462.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-constant": [{"name": "perl-constant", "version": "1.33", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Storable": [{"name": "perl-Storable", "version": "3.21", "release": "460.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "perl-overload": [{"name": "perl-overload", "version": "1.31", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-parent": [{"name": "perl-parent", "version": "0.238", "release": "460.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-vars": [{"name": "perl-vars", "version": "1.05", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Long": [{"name": "perl-Getopt-Long", "version": "2.52", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Carp": [{"name": "perl-Carp", "version": "1.50", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Exporter": [{"name": "perl-Exporter", "version": "5.74", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-NDBM_File": [{"name": "perl-NDBM_File", "version": "1.15", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-PathTools": [{"name": "perl-PathTools", "version": "3.78", "release": "461.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Encode": [{"name": "perl-Encode", "version": "3.08", "release": "462.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-libs": [{"name": "perl-libs", "version": "5.32.1", "release": "481.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-interpreter": [{"name": "perl-interpreter", "version": "5.32.1", "release": "481.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "aspell": [{"name": "aspell", "version": "0.60.8", "release": "8.el9", "epoch": 12, "arch": "x86_64", "source": "rpm"}], "tbb": [{"name": "tbb", "version": "2020.3", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dyninst": [{"name": "dyninst", "version": "12.1.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemtap-runtime": [{"name": "systemtap-runtime", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-headers": [{"name": "kernel-headers", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-headers": [{"name": "glibc-headers", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "strace": [{"name": "strace", "version": "5.18", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pkgconf-m4": [{"name": "pkgconf-m4", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libpkgconf": [{"name": "libpkgconf", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pkgconf": [{"name": "pkgconf", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pkgconf-pkg-config": [{"name": "pkgconf-pkg-config", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd-devel": [{"name": "libzstd-devel", "version": "1.5.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib-devel": [{"name": "zlib-devel", "version": "1.2.11", "release": "41.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf-devel": [{"name": "elfutils-libelf-devel", "version": "<<< 13118 1727204091.90253: stdout chunk (state=3): >>>0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-devel": [{"name": "glibc-devel", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt-devel": [{"name": "libxcrypt-devel", "version": "4.4.18", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gcc": [{"name": "gcc", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssl-devel": [{"name": "openssl-devel", "version": "3.2.2", "release": "6.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-devel": [{"name": "kernel-devel", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-devel": [{"name": "xz-devel", "version": "5.2.5", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-devel": [{"name": "elfutils-devel", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemtap-devel": [{"name": "systemtap-devel", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "efivar-libs": [{"name": "efivar-libs", "version": "38", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mokutil": [{"name": "mokutil", "version": "0.6.0", "release": "4.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "systemtap-client": [{"name": "systemtap-client", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemtap": [{"name": "systemtap", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "qa-tools": [{"name": "qa-tools", "version": "4.1", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libtirpc": [{"name": "libtirpc", "version": "1.3.3", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "git-core": [{"name": "git-core", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfsidmap": [{"name": "libnfsidmap", "version": "2.5.4", "release": "27.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "git-core-doc": [{"name": "git-core-doc", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpcbind": [{"name": "rpcbind", "version": "1.2.6", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wget": [{"name": "wget", "version": "1.21.1", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-lib": [{"name": "perl-lib", "version": "0.65", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Find": [{"name": "perl-File-Find", "version": "1.37", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Error": [{"name": "perl-Error", "version": "0.17029", "release": "7.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-DynaLoader": [{"name": "perl-DynaLoader", "version": "1.47", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-TermReadKey": [{"name": "perl-TermReadKey", "version": "2.38", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxslt": [{"name": "libxslt", "version": "1.1.34", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-lxml": [{"name": "python3-lxml", "version": "4.6.5", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gpm-libs": [{"name": "gpm-libs", "version": "1.20.7", "release": "29.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "emacs-filesystem": [{"name": "emacs-filesystem", "version": "27.2", "release": "10.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Git": [{"name": "perl-Git", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "git": [{"name": "git", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "yum-utils": [{"name": "yum-utils", "version": "4.3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "vim-filesystem": [{"name": "vim-filesystem", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "noarch", "source": "rpm"}], "vim-common": [{"name": "vim-common", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "time": [{"name": "time", "version": "1.9", "release": "18.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tar": [{"name": "tar", "version": "1.34", "release": "7.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "quota-nls": [{"name": "quota-nls", "version": "4.09", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "quota": [{"name": "quota", "version": "4.09", "release": "4.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-pyyaml": [{"name": "python3-pyyaml", "version": "5.4.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libev": [{"name": "libev", "version": "4.33", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto-libev": [{"name": "libverto-libev", "version": "0.3.2", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gssproxy": [{"name": "gssproxy", "version": "0.8.4", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nfs-utils": [{"name": "nfs-utils", "version": "2.5.4", "release": "27.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "bc": [{"name": "bc", "version": "1.07.1", "release": "14.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "beakerlib-redhat": [{"name": "beakerlib-redhat", "version": "1", "release": "35.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "beakerlib": [{"name": "beakerlib", "version": "1.29.3", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "restraint": [{"name": "restraint", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "restraint-rhts": [{"name": "restraint-rhts", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-enhanced": [{"name": "vim-enhanced", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "sssd-nfs-idmap": [{"name": "sssd-nfs-idmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsync": [{"name": "rsync", "version": "3.2.3", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools": [{"name": "python3-setuptools", "version": "53.0.0", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-distro": [{"name": "python3-distro", "version": "1.5.0", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-idna": [{"name": "python3-idna", "version": "2.10", "release": "7.el9.1", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-setools": [{"name": "python3-setools", "version": "4.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pytz": [{"name": "python3-pytz", "version": "2021.1", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-babel": [{"name": "python3-babel", "version": "2.9.1", "release": "2.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyserial": [{"name": "python3-pyserial", "version": "3.4", "release": "12.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyrsistent": [{"name": "python3-pyrsistent", "version": "0.17.3", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-prettytable": [{"name": "python3-prettytable", "version": "0.7.2", "release": "27.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-oauthlib": [{"name": "python3-oauthlib", "version": "3.1.1", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-netifaces": [{"name": "python3-netifaces", "version": "0.10.6", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-markupsafe": [{"name": "python3-markupsafe", "version": "1.1.1", "release": "12.el<<< 13118 1727204091.90284: stdout chunk (state=3): >>>9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jinja2": [{"name": "python3-jinja2", "version": "2.11.3", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libsemanage": [{"name": "python3-libsemanage", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jsonpointer": [{"name": "python3-jsonpointer", "version": "2.0", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonpatch": [{"name": "python3-jsonpatch", "version": "1.21", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-configobj": [{"name": "python3-configobj", "version": "5.0.6", "release": "25.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-audit": [{"name": "python3-audit", "version": "3.1.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-attrs": [{"name": "python3-attrs", "version": "20.3.0", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema": [{"name": "python3-jsonschema", "version": "3.2.0", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libmaxminddb": [{"name": "libmaxminddb", "version": "1.5.2", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "geolite2-country": [{"name": "geolite2-country", "version": "20191217", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "geolite2-city": [{"name": "geolite2-city", "version": "20191217", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "ipcalc": [{"name": "ipcalc", "version": "1.0.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdisk": [{"name": "gdisk", "version": "1.0.7", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "checkpolicy": [{"name": "checkpolicy", "version": "3.6", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-policycoreutils": [{"name": "python3-policycoreutils", "version": "3.6", "release": "2.1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pysocks": [{"name": "python3-pysocks", "version": "1.7.1", "release": "12.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-urllib3": [{"name": "python3-urllib3", "version": "1.26.5", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-chardet": [{"name": "python3-chardet", "version": "4.0.0", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-requests": [{"name": "python3-requests", "version": "2.25.1", "release": "8.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "dhcp-common": [{"name": "dhcp-common", "version": "4.4.2", "release": "19.b1.el9", "epoch": 12, "arch": "noarch", "source": "rpm"}], "dhcp-client": [{"name": "dhcp-client", "version": "4.4.2", "release": "19.b1.el9", "epoch": 12, "arch": "x86_64", "source": "rpm"}], "cloud-init": [{"name": "cloud-init", "version": "23.4", "release": "19.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "cloud-utils-growpart": [{"name": "cloud-utils-growpart", "version": "0.33", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jitterentropy": [{"name": "jitterentropy", "version": "3.5.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rng-tools": [{"name": "rng-tools", "version": "6.16", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip": [{"name": "python3-pip", "version": "21.3.1", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "hostapd": [{"name": "hostapd", "version": "2.10", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wpa_supplicant": [{"name": "wpa_supplicant", "version": "2.10", "release": "5.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dnsmasq": [{"name": "dnsmasq", "version": "2.85", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}]}}, "invocation": {"module_args": {"manager": ["auto"], "strategy": "first"}}} <<< 13118 1727204091.91858: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204091.91862: stderr chunk (state=3): >>><<< 13118 1727204091.91871: stdout chunk (state=3): >>><<< 13118 1727204091.91907: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"packages": {"libgcc": [{"name": "libgcc", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "linux-firmware-whence": [{"name": "linux-firmware-whence", "version": "20240905", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "tzdata": [{"name": "tzdata", "version": "2024a", "release": "2.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "linux-firmware": [{"name": "linux-firmware", "version": "20240905", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "gawk-all-langpacks": [{"name": "gawk-all-langpacks", "version": "5.1.0", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools-wheel": [{"name": "python3-setuptools-wheel", "version": "53.0.0", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "publicsuffix-list-dafsa": [{"name": "publicsuffix-list-dafsa", "version": "20210518", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "pcre2-syntax": [{"name": "pcre2-syntax", "version": "10.40", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "ncurses-base": [{"name": "ncurses-base", "version": "6.2", "release": "10.20210508.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libssh-config": [{"name": "libssh-config", "version": "0.10.4", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libreport-filesystem": [{"name": "libreport-filesystem", "version": "2.15.2", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf-data": [{"name": "dnf-data", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-misc": [{"name": "kbd-misc", "version": "2.4.0", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-legacy": [{"name": "kbd-legacy", "version": "2.4.0", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "hwdata": [{"name": "hwdata", "version": "0.348", "release": "9.15.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "fonts-filesystem": [{"name": "fonts-filesystem", "version": "2.0.5", "release": "7.el9.1", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dejavu-sans-fonts": [{"name": "dejavu-sans-fonts", "version": "2.37", "release": "18.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-core-font-en": [{"name": "langpacks-core-font-en", "version": "3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "firewalld-filesystem": [{"name": "firewalld-filesystem", "version": "1.3.4", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "coreutils-common": [{"name": "coreutils-common", "version": "8.32", "release": "36.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "centos-gpg-keys": [{"name": "centos-gpg-keys", "version": "9.0", "release": "26.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-repos": [{"name": "centos-stream-repos", "version": "9.0", "release": "26.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-release": [{"name": "centos-stream-release", "version": "9.0", "release": "26.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "setup": [{"name": "setup", "version": "2.13.7", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "filesystem": [{"name": "filesystem", "version": "3.16", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "basesystem": [{"name": "basesystem", "version": "11", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "glibc-gconv-extra": [{"name": "glibc-gconv-extra", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-langpack-en": [{"name": "glibc-langpack-en", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-common": [{"name": "glibc-common", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc": [{"name": "glibc", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses-libs": [{"name": "ncurses-libs", "version": "6.2", "release": "10.20210508.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bash": [{"name": "bash", "version": "5.1.8", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib": [{"name": "zlib", "version": "1.2.11", "release": "41.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-libs": [{"name": "xz-libs", "version": "5.2.5", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd": [{"name": "libzstd", "version": "1.5.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap": [{"name": "libcap", "version": "2.48", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "popt": [{"name": "popt", "version": "1.18", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bzip2-libs": [{"name": "bzip2-libs", "version": "1.0.8", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt": [{"name": "libxcrypt", "version": "4.4.18", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libuuid": [{"name": "libuuid", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sqlite-libs": [{"name": "sqlite-libs", "version": "3.34.1", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcom_err": [{"name": "libcom_err", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libstdc++": [{"name": "libstdc++", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmnl": [{"name": "libmnl", "version": "1.0.4", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf": [{"name": "elfutils-libelf", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxml2": [{"name": "libxml2", "version": "2.9.13", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies": [{"name": "crypto-policies", "version": "20240828", "release": "2.git626aa59.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "readline": [{"name": "readline", "version": "8.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "alternatives": [{"name": "alternatives", "version": "1.24", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng": [{"name": "libcap-ng", "version": "0.8.2", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit-libs": [{"name": "audit-libs", "version": "3.1.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libunistring": [{"name": "libunistring", "version": "0.9.10", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lua-libs": [{"name": "lua-libs", "version": "5.4.4", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libffi": [{"name": "libffi", "version": "3.4.2", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgpg-error": [{"name": "libgpg-error", "version": "1.42", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtalloc": [{"name": "libtalloc", "version": "2.4.2", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libidn2": [{"name": "libidn2", "version": "2.3.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gmp": [{"name": "gmp", "version": "6.2.0", "release": "13.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "keyutils-libs": [{"name": "keyutils-libs", "version": "1.6.3", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libattr": [{"name": "libattr", "version": "2.5.1", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libacl": [{"name": "libacl", "version": "2.3.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3": [{"name": "libnl3", "version": "3.9.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsepol": [{"name": "libsepol", "version": "3.6", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsmartcols": [{"name": "libsmartcols", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lz4-libs": [{"name": "lz4-libs", "version": "1.9.3", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre2": [{"name": "pcre2", "version": "10.40", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux": [{"name": "libselinux", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sed": [{"name": "sed", "version": "4.8", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "findutils": [{"name": "findutils", "version": "4.8.0", "release": "7.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libsemanage": [{"name": "libsemanage", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "shadow-utils": [{"name": "shadow-utils", "version": "4.9", "release": "9.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libtevent": [{"name": "libtevent", "version": "0.16.1", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgcrypt": [{"name": "libgcrypt", "version": "1.10.0", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file-libs": [{"name": "file-libs", "version": "5.39", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libedit": [{"name": "libedit", "version": "3.1", "release": "38.20210216cvs.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "psmisc": [{"name": "psmisc", "version": "23.4", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "expat": [{"name": "expat", "version": "2.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm-libs": [{"name": "gdbm-libs", "version": "1.23", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "jansson": [{"name": "jansson", "version": "2.14", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "json-c": [{"name": "json-c", "version": "0.14", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtasn1": [{"name": "libtasn1", "version": "4.16.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit": [{"name": "p11-kit", "version": "0.25.3", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtdb": [{"name": "libtdb", "version": "1.4.10", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit-trust": [{"name": "p11-kit-trust", "version": "0.25.3", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file": [{"name": "file", "version": "5.39", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libassuan": [{"name": "libassuan", "version": "2.5.5", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbpf": [{"name": "libbpf", "version": "1.4.0", "release": "1.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libnftnl": [{"name": "libnftnl", "version": "1.2.6", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse-libs": [{"name": "fuse-libs", "version": "2.9.9", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbasicobjects": [{"name": "libbasicobjects", "version": "0.1.1", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcollection": [{"name": "libcollection", "version": "0.7.0", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdb": [{"name": "libdb", "version": "5.3.28", "release": "54.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute": [{"name": "iproute", "version": "6.2.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdhash": [{"name": "libdhash", "version": "0.5.0", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgomp": [{"name": "libgomp", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libref_array": [{"name": "libref_array", "version": "0.1.5", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libseccomp": [{"name": "libseccomp", "version": "2.5.2", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsigsegv": [{"name": "libsigsegv", "version": "2.13", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_idmap": [{"name": "libsss_idmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lzo": [{"name": "lzo", "version": "2.10", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "numactl-libs": [{"name": "numactl-libs", "version": "2.0.18", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre": [{"name": "pcre", "version": "8.44", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grep": [{"name": "grep", "version": "3.6", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssl-libs": [{"name": "openssl-libs", "version": "3.2.2", "release": "6.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "coreutils": [{"name": "coreutils", "version": "8.32", "release": "36.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ca-certificates": [{"name": "ca-certificates", "version": "2024.2.69_v8.0.303", "release": "91.4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "systemd-libs": [{"name": "systemd-libs", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libblkid": [{"name": "libblkid", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-libs": [{"name": "dbus-libs", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libmount": [{"name": "libmount", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux-core": [{"name": "util-linux-core", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfdisk": [{"name": "libfdisk", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gzip": [{"name": "gzip", "version": "1.12", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod": [{"name": "kmod", "version": "28", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod-libs": [{"name": "kmod-libs", "version": "28", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib": [{"name": "cracklib", "version": "2.9.6", "release": "27.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "which": [{"name": "which", "version": "2.21", "release": "29.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib-dicts": [{"name": "cracklib-dicts", "version": "2.9.6", "release": "27.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-tools": [{"name": "dbus-tools", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "procps-ng": [{"name": "procps-ng", "version": "3.3.17", "release": "14.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssl": [{"name": "openssl", "version": "3.2.2", "release": "6.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libarchive": [{"name": "libarchive", "version": "3.5.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libevent": [{"name": "libevent", "version": "2.1.12", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_certmap": [{"name": "libsss_certmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz": [{"name": "xz", "version": "5.2.5", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "squashfs-tools": [{"name": "squashfs-tools", "version": "4.4", "release": "10.git1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcomps": [{"name": "libcomps", "version": "0.1.18", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libutempter": [{"name": "libutempter", "version": "1.2.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux-utils": [{"name": "libselinux-utils", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3-cli": [{"name": "libnl3-cli", "version": "3.9.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libteam": [{"name": "libteam", "version": "1.31", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "acl": [{"name": "acl", "version": "2.3.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-libs": [{"name": "gettext-libs", "version": "0.21", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext": [{"name": "gettext", "version": "0.21", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "attr": [{"name": "attr", "version": "2.5.1", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils": [{"name": "keyutils", "version": "1.6.3", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpfr": [{"name": "mpfr", "version": "4.1.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gawk": [{"name": "gawk", "version": "5.1.0", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpsl": [{"name": "libpsl", "version": "0.21.1", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libksba": [{"name": "libksba", "version": "1.5.1", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ethtool": [{"name": "ethtool", "version": "6.2", "release": "1.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "ipset-libs": [{"name": "ipset-libs", "version": "7.11", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ipset": [{"name": "ipset", "version": "7.11", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "groff-base": [{"name": "groff-base", "version": "1.22.4", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "snappy": [{"name": "snappy", "version": "1.1.8", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs-libs": [{"name": "e2fsprogs-libs", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libss": [{"name": "libss", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt-compat": [{"name": "libxcrypt-compat", "version": "4.4.18", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip-wheel": [{"name": "python3-pip-wheel", "version": "21.3.1", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python-unversioned-command": [{"name": "python-unversioned-command", "version": "3.9.19", "release": "8.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3": [{"name": "python3", "version": "3.9.19", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libs": [{"name": "python3-libs", "version": "3.9.19", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libcomps": [{"name": "python3-libcomps", "version": "0.1.18", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-six": [{"name": "python3-six", "version": "1.15.0", "release": "9.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dateutil": [{"name": "python3-dateutil", "version": "2.8.1", "release": "7.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "python3-systemd": [{"name": "python3-systemd", "version": "234", "release": "19.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng-python3": [{"name": "libcap-ng-python3", "version": "0.8.2", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pigz": [{"name": "pigz", "version": "2.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "hostname": [{"name": "hostname", "version": "3.23", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools-libs": [{"name": "kernel-tools-libs", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "less": [{"name": "less", "version": "590", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-rpm-macros": [{"name": "systemd-rpm-macros", "version": "252", "release": "47.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "c-ares": [{"name": "c-ares", "version": "1.19.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpio": [{"name": "cpio", "version": "2.13", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "diffutils": [{"name": "diffutils", "version": "3.7", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "inih": [{"name": "inih", "version": "49", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbrotli": [{"name": "libbrotli", "version": "1.0.9", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcbor": [{"name": "libcbor", "version": "0.7.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdaemon": [{"name": "libdaemon", "version": "0.14", "release": "23.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "teamd": [{"name": "teamd", "version": "1.31", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libeconf": [{"name": "libeconf", "version": "0.4.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpwquality": [{"name": "libpwquality", "version": "1.4.4", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pam": [{"name": "pam", "version": "1.5.1", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux": [{"name": "util-linux", "version": "2.37.4", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus": [{"name": "dbus", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "systemd-pam": [{"name": "systemd-pam", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd": [{"name": "systemd", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-common": [{"name": "dbus-common", "version": "1.12.20", "release": "8.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dbus-broker": [{"name": "dbus-broker", "version": "28", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-common": [{"name": "grub2-common", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "cronie-anacron": [{"name": "cronie-anacron", "version": "1.5.7", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cronie": [{"name": "cronie", "version": "1.5.7", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crontabs": [{"name": "crontabs", "version": "1.11", "release": "26.20190603git.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "openssh": [{"name": "openssh", "version": "8.7p1", "release": "43.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc-modules": [{"name": "grub2-pc-modules", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "authselect-libs": [{"name": "authselect-libs", "version": "1.2.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-libs": [{"name": "device-mapper-libs", "version": "1.02.198", "release": "2.el9", "epoch": 9, "arch": "x86_64", "source": "rpm"}], "device-mapper": [{"name": "device-mapper", "version": "1.02.198", "release": "2.el9", "epoch": 9, "arch": "x86_64", "source": "rpm"}], "grub2-tools-minimal": [{"name": "grub2-tools-minimal", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "cryptsetup-libs": [{"name": "cryptsetup-libs", "version": "2.7.2", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kpartx": [{"name": "kpartx", "version": "0.8.7", "release": "32.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-default-yama-scope": [{"name": "elfutils-default-yama-scope", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "elfutils-libs": [{"name": "elfutils-libs", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-service": [{"name": "initscripts-service", "version": "10.11.8", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iputils": [{"name": "iputils", "version": "20210202", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi": [{"name": "libkcapi", "version": "1.4.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hmaccalc": [{"name": "libkcapi-hmaccalc", "version": "1.4.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "logrotate": [{"name": "logrotate", "version": "3.18.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oddjob": [{"name": "oddjob", "version": "0.34.7", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oddjob-mkhomedir": [{"name": "oddjob-mkhomedir", "version": "0.34.7", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect": [{"name": "authselect", "version": "1.2.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kbd": [{"name": "kbd", "version": "2.4.0", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libndp": [{"name": "libndp", "version": "1.9", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfnetlink": [{"name": "libnfnetlink", "version": "1.0.1", "release": "21.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnetfilter_conntrack": [{"name": "libnetfilter_conntrack", "version": "1.0.9", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-libs": [{"name": "iptables-libs", "version": "1.8.10", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-nft": [{"name": "iptables-nft", "version": "1.8.10", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nftables": [{"name": "nftables", "version": "1.0.9", "release": "3.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-nftables": [{"name": "python3-nftables", "version": "1.0.9", "release": "3.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libnghttp2": [{"name": "libnghttp2", "version": "1.43.0", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpath_utils": [{"name": "libpath_utils", "version": "0.2.1", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libini_config": [{"name": "libini_config", "version": "1.3.1", "release": "53.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpipeline": [{"name": "libpipeline", "version": "1.5.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_nss_idmap": [{"name": "libsss_nss_idmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_sudo": [{"name": "libsss_sudo", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto": [{"name": "libverto", "version": "0.3.2", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "krb5-libs": [{"name": "krb5-libs", "version": "1.21.1", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-lib": [{"name": "cyrus-sasl-lib", "version": "2.1.27", "release": "21.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openldap": [{"name": "openldap", "version": "2.6.6", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libssh": [{"name": "libssh", "version": "0.10.4", "release": "13.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcurl": [{"name": "libcurl", "version": "7.76.1", "release": "31.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-debuginfod-client": [{"name": "elfutils-debuginfod-client", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils-gold": [{"name": "binutils-gold", "version": "2.35.2", "release": "54.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils": [{"name": "binutils", "version": "2.35.2", "release": "54.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tss": [{"name": "tpm2-tss", "version": "3.2.3", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-udev": [{"name": "systemd-udev", "version": "252", "release": "47.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut": [{"name": "dracut", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-modules-core": [{"name": "kernel-modules-core", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-core": [{"name": "kernel-core", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-modules": [{"name": "kernel-modules", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-squash": [{"name": "dracut-squash", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfido2": [{"name": "libfido2", "version": "1.13.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "os-prober": [{"name": "os-prober", "version": "1.77", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools": [{"name": "grub2-tools", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "ima-evm-utils": [{"name": "ima-evm-utils", "version": "1.5", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "curl": [{"name": "curl", "version": "7.76.1", "release": "31.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm": [{"name": "rpm", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-libs": [{"name": "rpm-libs", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grubby": [{"name": "grubby", "version": "8.40", "release": "63.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsolv": [{"name": "libsolv", "version": "0.7.24", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "policycoreutils": [{"name": "policycoreutils", "version": "3.6", "release": "2.1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "selinux-policy": [{"name": "selinux-policy", "version": "38.1.45", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "selinux-policy-targeted": [{"name": "selinux-policy-targeted", "version": "38.1.45", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpm-build-libs": [{"name": "rpm-build-libs", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-systemd-inhibit": [{"name": "rpm-plugin-systemd-inhibit", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-client": [{"name": "sssd-client", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libyaml": [{"name": "libyaml", "version": "0.2.5", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lmdb-libs": [{"name": "lmdb-libs", "version": "0.9.29", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libldb": [{"name": "libldb", "version": "2.9.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-common": [{"name": "sssd-common", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nettle": [{"name": "nettle", "version": "3.9.1", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnutls": [{"name": "gnutls", "version": "3.8.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glib2": [{"name": "glib2", "version": "2.68.4", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dbus": [{"name": "python3-dbus", "version": "1.2.18", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-libnm": [{"name": "NetworkManager-libnm", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "NetworkManager": [{"name": "NetworkManager", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libmodulemd": [{"name": "libmodulemd", "version": "2.13.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gobject-introspection": [{"name": "gobject-introspection", "version": "1.68.0", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gobject-base-noarch": [{"name": "python3-gobject-base-noarch", "version": "3.40.1", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-gobject-base": [{"name": "python3-gobject-base", "version": "3.40.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-firewall": [{"name": "python3-firewall", "version": "1.3.4", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libuser": [{"name": "libuser", "version": "0.63", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "npth": [{"name": "npth", "version": "1.6", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2": [{"name": "gnupg2", "version": "2.3.3", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gpgme": [{"name": "gpgme", "version": "1.15.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "librepo": [{"name": "librepo", "version": "1.14.5", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdnf": [{"name": "libdnf", "version": "0.69.0", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libdnf": [{"name": "python3-libdnf", "version": "0.69.0", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-hawkey": [{"name": "python3-hawkey", "version": "0.69.0", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gpg": [{"name": "python3-gpg", "version": "1.15.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sign-libs": [{"name": "rpm-sign-libs", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpm": [{"name": "python3-rpm", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dnf": [{"name": "python3-dnf", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf": [{"name": "dnf", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dnf-plugins-core": [{"name": "python3-dnf-plugins-core", "version": "4.3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "oniguruma": [{"name": "oniguruma", "version": "6.9.6", "release": "1.el9.6", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jq": [{"name": "jq", "version": "1.6", "release": "17.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-network": [{"name": "dracut-network", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pciutils-libs": [{"name": "pciutils-libs", "version": "3.7.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sg3_utils-libs": [{"name": "sg3_utils-libs", "version": "1.47", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "slang": [{"name": "slang", "version": "2.3.2", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "newt": [{"name": "newt", "version": "0.52.21", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "userspace-rcu": [{"name": "userspace-rcu", "version": "0.12.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libestr": [{"name": "libestr", "version": "0.1.11", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfastjson": [{"name": "libfastjson", "version": "0.99.9", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsyslog-logrotate": [{"name": "rsyslog-logrotate", "version": "8.2310.0", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsyslog": [{"name": "rsyslog", "version": "8.2310.0", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "liburing": [{"name": "liburing", "version": "2.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "langpacks-core-en": [{"name": "langpacks-core-en", "version": "3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-en": [{"name": "langpacks-en", "version": "3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "qemu-guest-agent": [{"name": "qemu-guest-agent", "version": "9.0.0", "release": "10.el9", "epoch": 17, "arch": "x86_64", "source": "rpm"}], "xfsprogs": [{"name": "xfsprogs", "version": "6.4.0", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-tui": [{"name": "NetworkManager-tui", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "sg3_utils": [{"name": "sg3_utils", "version": "1.47", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools": [{"name": "kernel-tools", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kexec-tools": [{"name": "kexec-tools", "version": "2.0.27", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dnf-plugins-core": [{"name": "dnf-plugins-core", "version": "4.3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "yum": [{"name": "yum", "version": "4.14.0", "release": "17.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "passwd": [{"name": "passwd", "version": "0.80", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "firewalld": [{"name": "firewalld", "version": "1.3.4", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "NetworkManager-team": [{"name": "NetworkManager-team", "version": "1.51.0", "release": "1.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "initscripts-rename-device": [{"name": "initscripts-rename-device", "version": "10.11.8", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "irqbalance": [{"name": "irqbalance", "version": "1.9.4", "release": "1.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "chrony": [{"name": "chrony", "version": "4.6", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-kcm": [{"name": "sssd-kcm", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-selinux": [{"name": "rpm-plugin-selinux", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies-scripts": [{"name": "crypto-policies-scripts", "version": "20240828", "release": "2.git626aa59.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpm-plugin-audit": [{"name": "rpm-plugin-audit", "version": "4.16.1.3", "release": "34.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc": [{"name": "grub2-pc", "version": "2.06", "release": "92.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "openssh-clients": [{"name": "openssh-clients", "version": "8.7p1", "release": "43.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel": [{"name": "kernel", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-config-rescue": [{"name": "dracut-config-rescue", "version": "057", "release": "70.git20240819.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo": [{"name": "sudo", "version": "1.9.5p2", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit": [{"name": "audit", "version": "3.1.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-server": [{"name": "openssh-server", "version": "8.7p1", "release": "43.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "man-db": [{"name": "man-db", "version": "2.9.3", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute-tc": [{"name": "iproute-tc", "version": "6.2.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect-compat": [{"name": "authselect-compat", "version": "1.2.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "parted": [{"name": "parted", "version": "3.5", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "microcode_ctl": [{"name": "microcode_ctl", "version": "20240531", "release": "1.el9", "epoch": 4, "arch": "noarch", "source": "rpm"}], "python3-libselinux": [{"name": "python3-libselinux", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs": [{"name": "e2fsprogs", "version": "1.46.5", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "prefixdevname": [{"name": "prefixdevname", "version": "0.1.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-minimal": [{"name": "vim-minimal", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "lshw": [{"name": "lshw", "version": "B.02.19.2", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses": [{"name": "ncurses", "version": "6.2", "release": "10.20210508.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsysfs": [{"name": "libsysfs", "version": "2.1.1", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lsscsi": [{"name": "lsscsi", "version": "0.32", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iwl100-firmware": [{"name": "iwl100-firmware", "version": "39.31.5.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl1000-firmware": [{"name": "iwl1000-firmware", "version": "39.31.5.1", "release": "146.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "iwl105-firmware": [{"name": "iwl105-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl135-firmware": [{"name": "iwl135-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl2000-firmware": [{"name": "iwl2000-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl2030-firmware": [{"name": "iwl2030-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl3160-firmware": [{"name": "iwl3160-firmware", "version": "25.30.13.0", "release": "146.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "iwl5000-firmware": [{"name": "iwl5000-firmware", "version": "8.83.5.1_1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl5150-firmware": [{"name": "iwl5150-firmware", "version": "8.24.2.2", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl6000g2a-firmware": [{"name": "iwl6000g2a-firmware", "version": "18.168.6.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl6050-firmware": [{"name": "iwl6050-firmware", "version": "41.28.5.1", "release": "146.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwl7260-firmware": [{"name": "iwl7260-firmware", "version": "25.30.13.0", "release": "146.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "rootfiles": [{"name": "rootfiles", "version": "8.1", "release": "31.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "gpg-pubkey": [{"name": "gpg-pubkey", "version": "3228467c", "release": "613798eb", "epoch": null, "arch": null, "source": "rpm"}, {"name": "gpg-pubkey", "version": "8483c65d", "release": "5ccc5b19", "epoch": null, "arch": null, "source": "rpm"}], "epel-release": [{"name": "epel-release", "version": "9", "release": "8.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "nspr": [{"name": "nspr", "version": "4.35.0", "release": "14.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "boost-system": [{"name": "boost-system", "version": "1.75.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-util": [{"name": "nss-util", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "make": [{"name": "make", "version": "4.3", "release": "8.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "m4": [{"name": "m4", "version": "1.4.19", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmpc": [{"name": "libmpc", "version": "1.2.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "unzip": [{"name": "unzip", "version": "6.0", "release": "57.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "avahi-libs": [{"name": "avahi-libs", "version": "0.8", "release": "21.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zip": [{"name": "zip", "version": "3.0", "release": "35.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpp": [{"name": "cpp", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bison": [{"name": "bison", "version": "3.7.4", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "flex": [{"name": "flex", "version": "2.6.4", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-softokn-freebl": [{"name": "nss-softokn-freebl", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-softokn": [{"name": "nss-softokn", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss": [{"name": "nss", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nss-sysinit": [{"name": "nss-sysinit", "version": "3.101.0", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "boost-filesystem": [{"name": "boost-filesystem", "version": "1.75.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "boost-thread": [{"name": "boost-thread", "version": "1.75.0", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Digest": [{"name": "perl-Digest", "version": "1.19", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Digest-MD5": [{"name": "perl-Digest-MD5", "version": "2.58", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-B": [{"name": "perl-B", "version": "1.80", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-FileHandle": [{"name": "perl-FileHandle", "version": "2.03", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Data-Dumper": [{"name": "perl-Data-Dumper", "version": "2.174", "release": "462.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-libnet": [{"name": "perl-libnet", "version": "3.13", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-base": [{"name": "perl-base", "version": "2.27", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-URI": [{"name": "perl-URI", "version": "5.09", "release": "3.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-AutoLoader": [{"name": "perl-AutoLoader", "version": "5.74", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Mozilla-CA": [{"name": "perl-Mozilla-CA", "version": "20200520", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-if": [{"name": "perl-if", "version": "0.60.800", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-IP": [{"name": "perl-IO-Socket-IP", "version": "0.41", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Time-Local": [{"name": "perl-Time-Local", "version": "1.300", "release": "7.el9", "epoch": 2, "arch": "noarch", "source": "rpm"}], "perl-File-Path": [{"name": "perl-File-Path", "version": "2.18", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Escapes": [{"name": "perl-Pod-Escapes", "version": "1.07", "release": "460.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Text-Tabs+Wrap": [{"name": "perl-Text-Tabs+Wrap", "version": "2013.0523", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-SSL": [{"name": "perl-IO-Socket-SSL", "version": "2.073", "release": "2.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Net-SSLeay": [{"name": "perl-Net-SSLeay", "version": "1.94", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Class-Struct": [{"name": "perl-Class-Struct", "version": "0.66", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-POSIX": [{"name": "perl-POSIX", "version": "1.94", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Term-ANSIColor": [{"name": "perl-Term-ANSIColor", "version": "5.01", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-IPC-Open3": [{"name": "perl-IPC-Open3", "version": "1.21", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-subs": [{"name": "perl-subs", "version": "1.03", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-Temp": [{"name": "perl-File-Temp", "version": "0.231.100", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Term-Cap": [{"name": "perl-Term-Cap", "version": "1.17", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Simple": [{"name": "perl-Pod-Simple", "version": "3.42", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-HTTP-Tiny": [{"name": "perl-HTTP-Tiny", "version": "0.076", "release": "462.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Socket": [{"name": "perl-Socket", "version": "2.031", "release": "4.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-SelectSaver": [{"name": "perl-SelectSaver", "version": "1.02", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Symbol": [{"name": "perl-Symbol", "version": "1.08", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-stat": [{"name": "perl-File-stat", "version": "1.09", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-podlators": [{"name": "perl-podlators", "version": "4.14", "release": "460.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Pod-Perldoc": [{"name": "perl-Pod-Perldoc", "version": "3.28.01", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Fcntl": [{"name": "perl-Fcntl", "version": "1.13", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Text-ParseWords": [{"name": "perl-Text-ParseWords", "version": "3.30", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-mro": [{"name": "perl-mro", "version": "1.23", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IO": [{"name": "perl-IO", "version": "1.43", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-overloading": [{"name": "perl-overloading", "version": "0.02", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Pod-Usage": [{"name": "perl-Pod-Usage", "version": "2.01", "release": "4.el9", "epoch": 4, "arch": "noarch", "source": "rpm"}], "perl-Errno": [{"name": "perl-Errno", "version": "1.30", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Basename": [{"name": "perl-File-Basename", "version": "2.85", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Std": [{"name": "perl-Getopt-Std", "version": "1.12", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-MIME-Base64": [{"name": "perl-MIME-Base64", "version": "3.16", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Scalar-List-Utils": [{"name": "perl-Scalar-List-Utils", "version": "1.56", "release": "462.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-constant": [{"name": "perl-constant", "version": "1.33", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Storable": [{"name": "perl-Storable", "version": "3.21", "release": "460.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "perl-overload": [{"name": "perl-overload", "version": "1.31", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-parent": [{"name": "perl-parent", "version": "0.238", "release": "460.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-vars": [{"name": "perl-vars", "version": "1.05", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Long": [{"name": "perl-Getopt-Long", "version": "2.52", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Carp": [{"name": "perl-Carp", "version": "1.50", "release": "460.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Exporter": [{"name": "perl-Exporter", "version": "5.74", "release": "461.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-NDBM_File": [{"name": "perl-NDBM_File", "version": "1.15", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-PathTools": [{"name": "perl-PathTools", "version": "3.78", "release": "461.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Encode": [{"name": "perl-Encode", "version": "3.08", "release": "462.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-libs": [{"name": "perl-libs", "version": "5.32.1", "release": "481.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-interpreter": [{"name": "perl-interpreter", "version": "5.32.1", "release": "481.el9", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "aspell": [{"name": "aspell", "version": "0.60.8", "release": "8.el9", "epoch": 12, "arch": "x86_64", "source": "rpm"}], "tbb": [{"name": "tbb", "version": "2020.3", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dyninst": [{"name": "dyninst", "version": "12.1.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemtap-runtime": [{"name": "systemtap-runtime", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-headers": [{"name": "kernel-headers", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-headers": [{"name": "glibc-headers", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "strace": [{"name": "strace", "version": "5.18", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pkgconf-m4": [{"name": "pkgconf-m4", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libpkgconf": [{"name": "libpkgconf", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pkgconf": [{"name": "pkgconf", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pkgconf-pkg-config": [{"name": "pkgconf-pkg-config", "version": "1.7.3", "release": "10.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd-devel": [{"name": "libzstd-devel", "version": "1.5.1", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib-devel": [{"name": "zlib-devel", "version": "1.2.11", "release": "41.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf-devel": [{"name": "elfutils-libelf-devel", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-devel": [{"name": "glibc-devel", "version": "2.34", "release": "125.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt-devel": [{"name": "libxcrypt-devel", "version": "4.4.18", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gcc": [{"name": "gcc", "version": "11.5.0", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssl-devel": [{"name": "openssl-devel", "version": "3.2.2", "release": "6.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-devel": [{"name": "kernel-devel", "version": "5.14.0", "release": "511.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-devel": [{"name": "xz-devel", "version": "5.2.5", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-devel": [{"name": "elfutils-devel", "version": "0.191", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemtap-devel": [{"name": "systemtap-devel", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "efivar-libs": [{"name": "efivar-libs", "version": "38", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mokutil": [{"name": "mokutil", "version": "0.6.0", "release": "4.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "systemtap-client": [{"name": "systemtap-client", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemtap": [{"name": "systemtap", "version": "5.1", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "qa-tools": [{"name": "qa-tools", "version": "4.1", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libtirpc": [{"name": "libtirpc", "version": "1.3.3", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "git-core": [{"name": "git-core", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfsidmap": [{"name": "libnfsidmap", "version": "2.5.4", "release": "27.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "git-core-doc": [{"name": "git-core-doc", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpcbind": [{"name": "rpcbind", "version": "1.2.6", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wget": [{"name": "wget", "version": "1.21.1", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-lib": [{"name": "perl-lib", "version": "0.65", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Find": [{"name": "perl-File-Find", "version": "1.37", "release": "481.el9", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Error": [{"name": "perl-Error", "version": "0.17029", "release": "7.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-DynaLoader": [{"name": "perl-DynaLoader", "version": "1.47", "release": "481.el9", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-TermReadKey": [{"name": "perl-TermReadKey", "version": "2.38", "release": "11.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxslt": [{"name": "libxslt", "version": "1.1.34", "release": "9.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-lxml": [{"name": "python3-lxml", "version": "4.6.5", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gpm-libs": [{"name": "gpm-libs", "version": "1.20.7", "release": "29.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "emacs-filesystem": [{"name": "emacs-filesystem", "version": "27.2", "release": "10.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Git": [{"name": "perl-Git", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "git": [{"name": "git", "version": "2.43.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "yum-utils": [{"name": "yum-utils", "version": "4.3.0", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "vim-filesystem": [{"name": "vim-filesystem", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "noarch", "source": "rpm"}], "vim-common": [{"name": "vim-common", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "time": [{"name": "time", "version": "1.9", "release": "18.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tar": [{"name": "tar", "version": "1.34", "release": "7.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "quota-nls": [{"name": "quota-nls", "version": "4.09", "release": "4.el9", "epoch": 1, "arch": "noarch", "source": "rpm"}], "quota": [{"name": "quota", "version": "4.09", "release": "4.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-pyyaml": [{"name": "python3-pyyaml", "version": "5.4.1", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libev": [{"name": "libev", "version": "4.33", "release": "6.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto-libev": [{"name": "libverto-libev", "version": "0.3.2", "release": "3.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gssproxy": [{"name": "gssproxy", "version": "0.8.4", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nfs-utils": [{"name": "nfs-utils", "version": "2.5.4", "release": "27.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "bc": [{"name": "bc", "version": "1.07.1", "release": "14.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "beakerlib-redhat": [{"name": "beakerlib-redhat", "version": "1", "release": "35.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "beakerlib": [{"name": "beakerlib", "version": "1.29.3", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "restraint": [{"name": "restraint", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "restraint-rhts": [{"name": "restraint-rhts", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-enhanced": [{"name": "vim-enhanced", "version": "8.2.2637", "release": "21.el9", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "sssd-nfs-idmap": [{"name": "sssd-nfs-idmap", "version": "2.9.5", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsync": [{"name": "rsync", "version": "3.2.3", "release": "20.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools": [{"name": "python3-setuptools", "version": "53.0.0", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-distro": [{"name": "python3-distro", "version": "1.5.0", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-idna": [{"name": "python3-idna", "version": "2.10", "release": "7.el9.1", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-setools": [{"name": "python3-setools", "version": "4.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pytz": [{"name": "python3-pytz", "version": "2021.1", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-babel": [{"name": "python3-babel", "version": "2.9.1", "release": "2.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyserial": [{"name": "python3-pyserial", "version": "3.4", "release": "12.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyrsistent": [{"name": "python3-pyrsistent", "version": "0.17.3", "release": "8.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-prettytable": [{"name": "python3-prettytable", "version": "0.7.2", "release": "27.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-oauthlib": [{"name": "python3-oauthlib", "version": "3.1.1", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-netifaces": [{"name": "python3-netifaces", "version": "0.10.6", "release": "15.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-markupsafe": [{"name": "python3-markupsafe", "version": "1.1.1", "release": "12.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jinja2": [{"name": "python3-jinja2", "version": "2.11.3", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libsemanage": [{"name": "python3-libsemanage", "version": "3.6", "release": "2.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jsonpointer": [{"name": "python3-jsonpointer", "version": "2.0", "release": "4.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonpatch": [{"name": "python3-jsonpatch", "version": "1.21", "release": "16.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-configobj": [{"name": "python3-configobj", "version": "5.0.6", "release": "25.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-audit": [{"name": "python3-audit", "version": "3.1.5", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-attrs": [{"name": "python3-attrs", "version": "20.3.0", "release": "7.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema": [{"name": "python3-jsonschema", "version": "3.2.0", "release": "13.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "libmaxminddb": [{"name": "libmaxminddb", "version": "1.5.2", "release": "4.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "geolite2-country": [{"name": "geolite2-country", "version": "20191217", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "geolite2-city": [{"name": "geolite2-city", "version": "20191217", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "ipcalc": [{"name": "ipcalc", "version": "1.0.0", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdisk": [{"name": "gdisk", "version": "1.0.7", "release": "5.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "checkpolicy": [{"name": "checkpolicy", "version": "3.6", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-policycoreutils": [{"name": "python3-policycoreutils", "version": "3.6", "release": "2.1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pysocks": [{"name": "python3-pysocks", "version": "1.7.1", "release": "12.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-urllib3": [{"name": "python3-urllib3", "version": "1.26.5", "release": "6.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-chardet": [{"name": "python3-chardet", "version": "4.0.0", "release": "5.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-requests": [{"name": "python3-requests", "version": "2.25.1", "release": "8.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "dhcp-common": [{"name": "dhcp-common", "version": "4.4.2", "release": "19.b1.el9", "epoch": 12, "arch": "noarch", "source": "rpm"}], "dhcp-client": [{"name": "dhcp-client", "version": "4.4.2", "release": "19.b1.el9", "epoch": 12, "arch": "x86_64", "source": "rpm"}], "cloud-init": [{"name": "cloud-init", "version": "23.4", "release": "19.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "cloud-utils-growpart": [{"name": "cloud-utils-growpart", "version": "0.33", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jitterentropy": [{"name": "jitterentropy", "version": "3.5.0", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rng-tools": [{"name": "rng-tools", "version": "6.16", "release": "7.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip": [{"name": "python3-pip", "version": "21.3.1", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "hostapd": [{"name": "hostapd", "version": "2.10", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wpa_supplicant": [{"name": "wpa_supplicant", "version": "2.10", "release": "5.el9", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dnsmasq": [{"name": "dnsmasq", "version": "2.85", "release": "16.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}]}}, "invocation": {"module_args": {"manager": ["auto"], "strategy": "first"}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204091.93665: done with _execute_module (package_facts, {'_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'package_facts', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204091.2973328-15447-256388152982611/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204091.93682: _low_level_execute_command(): starting 13118 1727204091.93685: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204091.2973328-15447-256388152982611/ > /dev/null 2>&1 && sleep 0' 13118 1727204091.94157: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204091.94161: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204091.94199: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.94211: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204091.94260: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204091.94278: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204091.94326: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204091.96150: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204091.96211: stderr chunk (state=3): >>><<< 13118 1727204091.96216: stdout chunk (state=3): >>><<< 13118 1727204091.96231: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204091.96239: handler run complete 13118 1727204091.96777: variable 'ansible_facts' from source: unknown 13118 1727204091.97093: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204091.98507: variable 'ansible_facts' from source: unknown 13118 1727204091.99017: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204091.99734: attempt loop complete, returning result 13118 1727204091.99746: _execute() done 13118 1727204091.99750: dumping result to json 13118 1727204091.99955: done dumping result, returning 13118 1727204091.99965: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check which packages are installed [0affcd87-79f5-56a3-0a64-000000000497] 13118 1727204091.99972: sending task result for task 0affcd87-79f5-56a3-0a64-000000000497 13118 1727204092.01398: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000497 13118 1727204092.01401: WORKER PROCESS EXITING ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13118 1727204092.01486: no more pending results, returning what we have 13118 1727204092.01488: results queue empty 13118 1727204092.01489: checking for any_errors_fatal 13118 1727204092.01492: done checking for any_errors_fatal 13118 1727204092.01493: checking for max_fail_percentage 13118 1727204092.01494: done checking for max_fail_percentage 13118 1727204092.01494: checking to see if all hosts have failed and the running result is not ok 13118 1727204092.01495: done checking to see if all hosts have failed 13118 1727204092.01495: getting the remaining hosts for this loop 13118 1727204092.01496: done getting the remaining hosts for this loop 13118 1727204092.01499: getting the next task for host managed-node2 13118 1727204092.01504: done getting next task for host managed-node2 13118 1727204092.01507: ^ task is: TASK: fedora.linux_system_roles.network : Print network provider 13118 1727204092.01510: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204092.01517: getting variables 13118 1727204092.01518: in VariableManager get_vars() 13118 1727204092.01545: Calling all_inventory to load vars for managed-node2 13118 1727204092.01547: Calling groups_inventory to load vars for managed-node2 13118 1727204092.01548: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204092.01555: Calling all_plugins_play to load vars for managed-node2 13118 1727204092.01557: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204092.01558: Calling groups_plugins_play to load vars for managed-node2 13118 1727204092.02313: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204092.03220: done with get_vars() 13118 1727204092.03239: done getting variables 13118 1727204092.03287: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Print network provider] ************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:7 Tuesday 24 September 2024 14:54:52 -0400 (0:00:00.802) 0:00:29.333 ***** 13118 1727204092.03320: entering _queue_task() for managed-node2/debug 13118 1727204092.03562: worker is 1 (out of 1 available) 13118 1727204092.03577: exiting _queue_task() for managed-node2/debug 13118 1727204092.03590: done queuing things up, now waiting for results queue to drain 13118 1727204092.03591: waiting for pending results... 13118 1727204092.03782: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Print network provider 13118 1727204092.03886: in run() - task 0affcd87-79f5-56a3-0a64-00000000007d 13118 1727204092.03897: variable 'ansible_search_path' from source: unknown 13118 1727204092.03901: variable 'ansible_search_path' from source: unknown 13118 1727204092.03931: calling self._execute() 13118 1727204092.04008: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204092.04012: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204092.04020: variable 'omit' from source: magic vars 13118 1727204092.04298: variable 'ansible_distribution_major_version' from source: facts 13118 1727204092.04311: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204092.04318: variable 'omit' from source: magic vars 13118 1727204092.04369: variable 'omit' from source: magic vars 13118 1727204092.04449: variable 'network_provider' from source: set_fact 13118 1727204092.04463: variable 'omit' from source: magic vars 13118 1727204092.04504: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204092.04533: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204092.04552: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204092.04567: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204092.04579: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204092.04603: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204092.04606: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204092.04610: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204092.04679: Set connection var ansible_timeout to 10 13118 1727204092.04690: Set connection var ansible_pipelining to False 13118 1727204092.04692: Set connection var ansible_connection to ssh 13118 1727204092.04697: Set connection var ansible_shell_executable to /bin/sh 13118 1727204092.04703: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204092.04705: Set connection var ansible_shell_type to sh 13118 1727204092.04724: variable 'ansible_shell_executable' from source: unknown 13118 1727204092.04729: variable 'ansible_connection' from source: unknown 13118 1727204092.04731: variable 'ansible_module_compression' from source: unknown 13118 1727204092.04733: variable 'ansible_shell_type' from source: unknown 13118 1727204092.04736: variable 'ansible_shell_executable' from source: unknown 13118 1727204092.04738: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204092.04742: variable 'ansible_pipelining' from source: unknown 13118 1727204092.04748: variable 'ansible_timeout' from source: unknown 13118 1727204092.04752: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204092.04856: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204092.04868: variable 'omit' from source: magic vars 13118 1727204092.04872: starting attempt loop 13118 1727204092.04876: running the handler 13118 1727204092.04913: handler run complete 13118 1727204092.04923: attempt loop complete, returning result 13118 1727204092.04926: _execute() done 13118 1727204092.04929: dumping result to json 13118 1727204092.04934: done dumping result, returning 13118 1727204092.04939: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Print network provider [0affcd87-79f5-56a3-0a64-00000000007d] 13118 1727204092.04944: sending task result for task 0affcd87-79f5-56a3-0a64-00000000007d 13118 1727204092.05027: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000007d 13118 1727204092.05032: WORKER PROCESS EXITING ok: [managed-node2] => {} MSG: Using network provider: nm 13118 1727204092.05097: no more pending results, returning what we have 13118 1727204092.05101: results queue empty 13118 1727204092.05102: checking for any_errors_fatal 13118 1727204092.05113: done checking for any_errors_fatal 13118 1727204092.05113: checking for max_fail_percentage 13118 1727204092.05115: done checking for max_fail_percentage 13118 1727204092.05116: checking to see if all hosts have failed and the running result is not ok 13118 1727204092.05117: done checking to see if all hosts have failed 13118 1727204092.05117: getting the remaining hosts for this loop 13118 1727204092.05119: done getting the remaining hosts for this loop 13118 1727204092.05123: getting the next task for host managed-node2 13118 1727204092.05134: done getting next task for host managed-node2 13118 1727204092.05138: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider 13118 1727204092.05142: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204092.05154: getting variables 13118 1727204092.05156: in VariableManager get_vars() 13118 1727204092.05199: Calling all_inventory to load vars for managed-node2 13118 1727204092.05202: Calling groups_inventory to load vars for managed-node2 13118 1727204092.05204: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204092.05214: Calling all_plugins_play to load vars for managed-node2 13118 1727204092.05216: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204092.05219: Calling groups_plugins_play to load vars for managed-node2 13118 1727204092.06010: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204092.06934: done with get_vars() 13118 1727204092.06957: done getting variables 13118 1727204092.07004: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:11 Tuesday 24 September 2024 14:54:52 -0400 (0:00:00.037) 0:00:29.370 ***** 13118 1727204092.07035: entering _queue_task() for managed-node2/fail 13118 1727204092.07308: worker is 1 (out of 1 available) 13118 1727204092.07321: exiting _queue_task() for managed-node2/fail 13118 1727204092.07336: done queuing things up, now waiting for results queue to drain 13118 1727204092.07337: waiting for pending results... 13118 1727204092.07515: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider 13118 1727204092.07622: in run() - task 0affcd87-79f5-56a3-0a64-00000000007e 13118 1727204092.07635: variable 'ansible_search_path' from source: unknown 13118 1727204092.07640: variable 'ansible_search_path' from source: unknown 13118 1727204092.07668: calling self._execute() 13118 1727204092.07740: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204092.07744: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204092.07752: variable 'omit' from source: magic vars 13118 1727204092.08027: variable 'ansible_distribution_major_version' from source: facts 13118 1727204092.08038: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204092.08124: variable 'network_state' from source: role '' defaults 13118 1727204092.08135: Evaluated conditional (network_state != {}): False 13118 1727204092.08139: when evaluation is False, skipping this task 13118 1727204092.08141: _execute() done 13118 1727204092.08144: dumping result to json 13118 1727204092.08146: done dumping result, returning 13118 1727204092.08151: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider [0affcd87-79f5-56a3-0a64-00000000007e] 13118 1727204092.08157: sending task result for task 0affcd87-79f5-56a3-0a64-00000000007e 13118 1727204092.08247: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000007e 13118 1727204092.08250: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13118 1727204092.08297: no more pending results, returning what we have 13118 1727204092.08302: results queue empty 13118 1727204092.08303: checking for any_errors_fatal 13118 1727204092.08309: done checking for any_errors_fatal 13118 1727204092.08310: checking for max_fail_percentage 13118 1727204092.08312: done checking for max_fail_percentage 13118 1727204092.08312: checking to see if all hosts have failed and the running result is not ok 13118 1727204092.08313: done checking to see if all hosts have failed 13118 1727204092.08314: getting the remaining hosts for this loop 13118 1727204092.08315: done getting the remaining hosts for this loop 13118 1727204092.08319: getting the next task for host managed-node2 13118 1727204092.08326: done getting next task for host managed-node2 13118 1727204092.08332: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 13118 1727204092.08337: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204092.08357: getting variables 13118 1727204092.08363: in VariableManager get_vars() 13118 1727204092.08401: Calling all_inventory to load vars for managed-node2 13118 1727204092.08404: Calling groups_inventory to load vars for managed-node2 13118 1727204092.08406: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204092.08416: Calling all_plugins_play to load vars for managed-node2 13118 1727204092.08418: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204092.08421: Calling groups_plugins_play to load vars for managed-node2 13118 1727204092.13887: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204092.16606: done with get_vars() 13118 1727204092.16649: done getting variables 13118 1727204092.16725: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:18 Tuesday 24 September 2024 14:54:52 -0400 (0:00:00.097) 0:00:29.467 ***** 13118 1727204092.16771: entering _queue_task() for managed-node2/fail 13118 1727204092.17236: worker is 1 (out of 1 available) 13118 1727204092.17250: exiting _queue_task() for managed-node2/fail 13118 1727204092.17262: done queuing things up, now waiting for results queue to drain 13118 1727204092.17266: waiting for pending results... 13118 1727204092.17663: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 13118 1727204092.17897: in run() - task 0affcd87-79f5-56a3-0a64-00000000007f 13118 1727204092.17919: variable 'ansible_search_path' from source: unknown 13118 1727204092.17937: variable 'ansible_search_path' from source: unknown 13118 1727204092.18006: calling self._execute() 13118 1727204092.18136: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204092.18156: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204092.18211: variable 'omit' from source: magic vars 13118 1727204092.18685: variable 'ansible_distribution_major_version' from source: facts 13118 1727204092.18712: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204092.18882: variable 'network_state' from source: role '' defaults 13118 1727204092.18899: Evaluated conditional (network_state != {}): False 13118 1727204092.18907: when evaluation is False, skipping this task 13118 1727204092.18916: _execute() done 13118 1727204092.18928: dumping result to json 13118 1727204092.18941: done dumping result, returning 13118 1727204092.18954: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 [0affcd87-79f5-56a3-0a64-00000000007f] 13118 1727204092.18968: sending task result for task 0affcd87-79f5-56a3-0a64-00000000007f skipping: [managed-node2] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13118 1727204092.19162: no more pending results, returning what we have 13118 1727204092.19169: results queue empty 13118 1727204092.19170: checking for any_errors_fatal 13118 1727204092.19180: done checking for any_errors_fatal 13118 1727204092.19181: checking for max_fail_percentage 13118 1727204092.19183: done checking for max_fail_percentage 13118 1727204092.19184: checking to see if all hosts have failed and the running result is not ok 13118 1727204092.19185: done checking to see if all hosts have failed 13118 1727204092.19186: getting the remaining hosts for this loop 13118 1727204092.19187: done getting the remaining hosts for this loop 13118 1727204092.19192: getting the next task for host managed-node2 13118 1727204092.19202: done getting next task for host managed-node2 13118 1727204092.19208: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later 13118 1727204092.19213: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204092.19241: getting variables 13118 1727204092.19243: in VariableManager get_vars() 13118 1727204092.19320: Calling all_inventory to load vars for managed-node2 13118 1727204092.19324: Calling groups_inventory to load vars for managed-node2 13118 1727204092.19326: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204092.19341: Calling all_plugins_play to load vars for managed-node2 13118 1727204092.19344: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204092.19347: Calling groups_plugins_play to load vars for managed-node2 13118 1727204092.20687: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000007f 13118 1727204092.20693: WORKER PROCESS EXITING 13118 1727204092.21932: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204092.23871: done with get_vars() 13118 1727204092.23907: done getting variables 13118 1727204092.24049: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:25 Tuesday 24 September 2024 14:54:52 -0400 (0:00:00.073) 0:00:29.541 ***** 13118 1727204092.24103: entering _queue_task() for managed-node2/fail 13118 1727204092.25211: worker is 1 (out of 1 available) 13118 1727204092.25226: exiting _queue_task() for managed-node2/fail 13118 1727204092.25263: done queuing things up, now waiting for results queue to drain 13118 1727204092.25266: waiting for pending results... 13118 1727204092.25625: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later 13118 1727204092.25863: in run() - task 0affcd87-79f5-56a3-0a64-000000000080 13118 1727204092.25966: variable 'ansible_search_path' from source: unknown 13118 1727204092.25977: variable 'ansible_search_path' from source: unknown 13118 1727204092.26022: calling self._execute() 13118 1727204092.26176: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204092.26189: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204092.26224: variable 'omit' from source: magic vars 13118 1727204092.26826: variable 'ansible_distribution_major_version' from source: facts 13118 1727204092.26850: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204092.27087: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204092.31539: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204092.31825: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204092.31900: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204092.31948: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204092.32069: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204092.32161: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204092.32201: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204092.32238: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204092.32294: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204092.32315: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204092.32429: variable 'ansible_distribution_major_version' from source: facts 13118 1727204092.32454: Evaluated conditional (ansible_distribution_major_version | int > 9): False 13118 1727204092.32463: when evaluation is False, skipping this task 13118 1727204092.32476: _execute() done 13118 1727204092.32484: dumping result to json 13118 1727204092.32492: done dumping result, returning 13118 1727204092.32504: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later [0affcd87-79f5-56a3-0a64-000000000080] 13118 1727204092.32514: sending task result for task 0affcd87-79f5-56a3-0a64-000000000080 skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_distribution_major_version | int > 9", "skip_reason": "Conditional result was False" } 13118 1727204092.32681: no more pending results, returning what we have 13118 1727204092.32686: results queue empty 13118 1727204092.32687: checking for any_errors_fatal 13118 1727204092.32694: done checking for any_errors_fatal 13118 1727204092.32695: checking for max_fail_percentage 13118 1727204092.32696: done checking for max_fail_percentage 13118 1727204092.32697: checking to see if all hosts have failed and the running result is not ok 13118 1727204092.32699: done checking to see if all hosts have failed 13118 1727204092.32699: getting the remaining hosts for this loop 13118 1727204092.32701: done getting the remaining hosts for this loop 13118 1727204092.32705: getting the next task for host managed-node2 13118 1727204092.32714: done getting next task for host managed-node2 13118 1727204092.32718: ^ task is: TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces 13118 1727204092.32723: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204092.32744: getting variables 13118 1727204092.32746: in VariableManager get_vars() 13118 1727204092.32791: Calling all_inventory to load vars for managed-node2 13118 1727204092.32794: Calling groups_inventory to load vars for managed-node2 13118 1727204092.32797: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204092.32809: Calling all_plugins_play to load vars for managed-node2 13118 1727204092.32813: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204092.32817: Calling groups_plugins_play to load vars for managed-node2 13118 1727204092.34251: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000080 13118 1727204092.34255: WORKER PROCESS EXITING 13118 1727204092.34987: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204092.37170: done with get_vars() 13118 1727204092.37203: done getting variables 13118 1727204092.37384: Loading ActionModule 'dnf' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/dnf.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:36 Tuesday 24 September 2024 14:54:52 -0400 (0:00:00.133) 0:00:29.674 ***** 13118 1727204092.37421: entering _queue_task() for managed-node2/dnf 13118 1727204092.37859: worker is 1 (out of 1 available) 13118 1727204092.37873: exiting _queue_task() for managed-node2/dnf 13118 1727204092.37886: done queuing things up, now waiting for results queue to drain 13118 1727204092.37887: waiting for pending results... 13118 1727204092.38761: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces 13118 1727204092.39144: in run() - task 0affcd87-79f5-56a3-0a64-000000000081 13118 1727204092.39157: variable 'ansible_search_path' from source: unknown 13118 1727204092.39162: variable 'ansible_search_path' from source: unknown 13118 1727204092.39385: calling self._execute() 13118 1727204092.39491: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204092.39496: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204092.39505: variable 'omit' from source: magic vars 13118 1727204092.40352: variable 'ansible_distribution_major_version' from source: facts 13118 1727204092.40368: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204092.40873: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204092.43880: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204092.44016: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204092.44063: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204092.44107: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204092.44142: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204092.44233: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204092.44272: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204092.44311: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204092.44369: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204092.44393: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204092.44606: variable 'ansible_distribution' from source: facts 13118 1727204092.44617: variable 'ansible_distribution_major_version' from source: facts 13118 1727204092.44647: Evaluated conditional (ansible_distribution == 'Fedora' or ansible_distribution_major_version | int > 7): True 13118 1727204092.44782: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204092.44929: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204092.45051: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204092.45090: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204092.45138: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204092.45160: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204092.45214: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204092.45246: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204092.45281: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204092.45332: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204092.45353: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204092.45403: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204092.45437: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204092.45470: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204092.45519: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204092.45544: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204092.45724: variable 'network_connections' from source: task vars 13118 1727204092.45749: variable 'port2_profile' from source: play vars 13118 1727204092.45824: variable 'port2_profile' from source: play vars 13118 1727204092.45847: variable 'port1_profile' from source: play vars 13118 1727204092.45913: variable 'port1_profile' from source: play vars 13118 1727204092.45927: variable 'controller_profile' from source: play vars 13118 1727204092.45999: variable 'controller_profile' from source: play vars 13118 1727204092.46197: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204092.46503: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204092.46554: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204092.46594: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204092.46636: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204092.46791: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204092.46833: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204092.46868: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204092.46902: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204092.46960: variable '__network_team_connections_defined' from source: role '' defaults 13118 1727204092.47373: variable 'network_connections' from source: task vars 13118 1727204092.47383: variable 'port2_profile' from source: play vars 13118 1727204092.47484: variable 'port2_profile' from source: play vars 13118 1727204092.47582: variable 'port1_profile' from source: play vars 13118 1727204092.47654: variable 'port1_profile' from source: play vars 13118 1727204092.47670: variable 'controller_profile' from source: play vars 13118 1727204092.47736: variable 'controller_profile' from source: play vars 13118 1727204092.47773: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13118 1727204092.47782: when evaluation is False, skipping this task 13118 1727204092.47790: _execute() done 13118 1727204092.47797: dumping result to json 13118 1727204092.47804: done dumping result, returning 13118 1727204092.47816: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces [0affcd87-79f5-56a3-0a64-000000000081] 13118 1727204092.47827: sending task result for task 0affcd87-79f5-56a3-0a64-000000000081 skipping: [managed-node2] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13118 1727204092.48011: no more pending results, returning what we have 13118 1727204092.48016: results queue empty 13118 1727204092.48017: checking for any_errors_fatal 13118 1727204092.48025: done checking for any_errors_fatal 13118 1727204092.48026: checking for max_fail_percentage 13118 1727204092.48028: done checking for max_fail_percentage 13118 1727204092.48031: checking to see if all hosts have failed and the running result is not ok 13118 1727204092.48032: done checking to see if all hosts have failed 13118 1727204092.48033: getting the remaining hosts for this loop 13118 1727204092.48034: done getting the remaining hosts for this loop 13118 1727204092.48039: getting the next task for host managed-node2 13118 1727204092.48048: done getting next task for host managed-node2 13118 1727204092.48052: ^ task is: TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces 13118 1727204092.48056: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204092.48078: getting variables 13118 1727204092.48080: in VariableManager get_vars() 13118 1727204092.48122: Calling all_inventory to load vars for managed-node2 13118 1727204092.48125: Calling groups_inventory to load vars for managed-node2 13118 1727204092.48128: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204092.48142: Calling all_plugins_play to load vars for managed-node2 13118 1727204092.48146: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204092.48149: Calling groups_plugins_play to load vars for managed-node2 13118 1727204092.49726: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000081 13118 1727204092.49732: WORKER PROCESS EXITING 13118 1727204092.50345: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204092.55194: done with get_vars() 13118 1727204092.55232: done getting variables redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf 13118 1727204092.55314: Loading ActionModule 'ansible_collections.ansible.builtin.plugins.action.dnf' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/dnf.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:48 Tuesday 24 September 2024 14:54:52 -0400 (0:00:00.179) 0:00:29.853 ***** 13118 1727204092.55353: entering _queue_task() for managed-node2/yum 13118 1727204092.55706: worker is 1 (out of 1 available) 13118 1727204092.55719: exiting _queue_task() for managed-node2/yum 13118 1727204092.55734: done queuing things up, now waiting for results queue to drain 13118 1727204092.55735: waiting for pending results... 13118 1727204092.57822: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces 13118 1727204092.58435: in run() - task 0affcd87-79f5-56a3-0a64-000000000082 13118 1727204092.58445: variable 'ansible_search_path' from source: unknown 13118 1727204092.58449: variable 'ansible_search_path' from source: unknown 13118 1727204092.58688: calling self._execute() 13118 1727204092.59020: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204092.59024: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204092.59035: variable 'omit' from source: magic vars 13118 1727204092.60141: variable 'ansible_distribution_major_version' from source: facts 13118 1727204092.60153: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204092.60581: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204092.66026: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204092.66224: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204092.66375: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204092.66412: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204092.66439: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204092.66636: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204092.66665: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204092.66693: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204092.66847: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204092.66863: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204092.67079: variable 'ansible_distribution_major_version' from source: facts 13118 1727204092.67095: Evaluated conditional (ansible_distribution_major_version | int < 8): False 13118 1727204092.67098: when evaluation is False, skipping this task 13118 1727204092.67101: _execute() done 13118 1727204092.67104: dumping result to json 13118 1727204092.67109: done dumping result, returning 13118 1727204092.67116: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces [0affcd87-79f5-56a3-0a64-000000000082] 13118 1727204092.67122: sending task result for task 0affcd87-79f5-56a3-0a64-000000000082 13118 1727204092.67711: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000082 13118 1727204092.67714: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_distribution_major_version | int < 8", "skip_reason": "Conditional result was False" } 13118 1727204092.67756: no more pending results, returning what we have 13118 1727204092.67760: results queue empty 13118 1727204092.67761: checking for any_errors_fatal 13118 1727204092.67769: done checking for any_errors_fatal 13118 1727204092.67769: checking for max_fail_percentage 13118 1727204092.67772: done checking for max_fail_percentage 13118 1727204092.67773: checking to see if all hosts have failed and the running result is not ok 13118 1727204092.67774: done checking to see if all hosts have failed 13118 1727204092.67775: getting the remaining hosts for this loop 13118 1727204092.67776: done getting the remaining hosts for this loop 13118 1727204092.67779: getting the next task for host managed-node2 13118 1727204092.67787: done getting next task for host managed-node2 13118 1727204092.67791: ^ task is: TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces 13118 1727204092.67795: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204092.67811: getting variables 13118 1727204092.67812: in VariableManager get_vars() 13118 1727204092.67854: Calling all_inventory to load vars for managed-node2 13118 1727204092.67857: Calling groups_inventory to load vars for managed-node2 13118 1727204092.67859: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204092.67872: Calling all_plugins_play to load vars for managed-node2 13118 1727204092.67875: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204092.67878: Calling groups_plugins_play to load vars for managed-node2 13118 1727204092.70096: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204092.75161: done with get_vars() 13118 1727204092.75296: done getting variables 13118 1727204092.75359: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:60 Tuesday 24 September 2024 14:54:52 -0400 (0:00:00.200) 0:00:30.053 ***** 13118 1727204092.75399: entering _queue_task() for managed-node2/fail 13118 1727204092.75743: worker is 1 (out of 1 available) 13118 1727204092.75756: exiting _queue_task() for managed-node2/fail 13118 1727204092.75770: done queuing things up, now waiting for results queue to drain 13118 1727204092.75772: waiting for pending results... 13118 1727204092.77251: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces 13118 1727204092.77836: in run() - task 0affcd87-79f5-56a3-0a64-000000000083 13118 1727204092.78085: variable 'ansible_search_path' from source: unknown 13118 1727204092.78089: variable 'ansible_search_path' from source: unknown 13118 1727204092.78132: calling self._execute() 13118 1727204092.78335: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204092.78340: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204092.78349: variable 'omit' from source: magic vars 13118 1727204092.79160: variable 'ansible_distribution_major_version' from source: facts 13118 1727204092.79175: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204092.79300: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204092.79512: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204092.83309: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204092.83750: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204092.83794: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204092.83826: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204092.83853: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204092.83937: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204092.83962: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204092.83994: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204092.84035: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204092.84049: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204092.84100: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204092.84122: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204092.84147: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204092.84189: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204092.84207: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204092.84248: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204092.84272: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204092.84296: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204092.84338: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204092.84353: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204092.84535: variable 'network_connections' from source: task vars 13118 1727204092.84548: variable 'port2_profile' from source: play vars 13118 1727204092.84621: variable 'port2_profile' from source: play vars 13118 1727204092.84636: variable 'port1_profile' from source: play vars 13118 1727204092.84700: variable 'port1_profile' from source: play vars 13118 1727204092.84708: variable 'controller_profile' from source: play vars 13118 1727204092.84776: variable 'controller_profile' from source: play vars 13118 1727204092.84850: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204092.85018: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204092.85053: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204092.85090: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204092.85135: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204092.85181: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204092.85203: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204092.85229: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204092.85255: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204092.85313: variable '__network_team_connections_defined' from source: role '' defaults 13118 1727204092.85590: variable 'network_connections' from source: task vars 13118 1727204092.85593: variable 'port2_profile' from source: play vars 13118 1727204092.85768: variable 'port2_profile' from source: play vars 13118 1727204092.85776: variable 'port1_profile' from source: play vars 13118 1727204092.85948: variable 'port1_profile' from source: play vars 13118 1727204092.85955: variable 'controller_profile' from source: play vars 13118 1727204092.86087: variable 'controller_profile' from source: play vars 13118 1727204092.86111: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13118 1727204092.86124: when evaluation is False, skipping this task 13118 1727204092.86126: _execute() done 13118 1727204092.86132: dumping result to json 13118 1727204092.86134: done dumping result, returning 13118 1727204092.86137: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces [0affcd87-79f5-56a3-0a64-000000000083] 13118 1727204092.86138: sending task result for task 0affcd87-79f5-56a3-0a64-000000000083 13118 1727204092.86350: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000083 13118 1727204092.86353: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13118 1727204092.86424: no more pending results, returning what we have 13118 1727204092.86429: results queue empty 13118 1727204092.86432: checking for any_errors_fatal 13118 1727204092.86438: done checking for any_errors_fatal 13118 1727204092.86439: checking for max_fail_percentage 13118 1727204092.86441: done checking for max_fail_percentage 13118 1727204092.86442: checking to see if all hosts have failed and the running result is not ok 13118 1727204092.86443: done checking to see if all hosts have failed 13118 1727204092.86444: getting the remaining hosts for this loop 13118 1727204092.86445: done getting the remaining hosts for this loop 13118 1727204092.86449: getting the next task for host managed-node2 13118 1727204092.86458: done getting next task for host managed-node2 13118 1727204092.86462: ^ task is: TASK: fedora.linux_system_roles.network : Install packages 13118 1727204092.86469: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204092.86486: getting variables 13118 1727204092.86488: in VariableManager get_vars() 13118 1727204092.86533: Calling all_inventory to load vars for managed-node2 13118 1727204092.86535: Calling groups_inventory to load vars for managed-node2 13118 1727204092.86537: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204092.86547: Calling all_plugins_play to load vars for managed-node2 13118 1727204092.86549: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204092.86552: Calling groups_plugins_play to load vars for managed-node2 13118 1727204092.89645: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204092.93554: done with get_vars() 13118 1727204092.93593: done getting variables 13118 1727204092.93661: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install packages] ******************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:73 Tuesday 24 September 2024 14:54:52 -0400 (0:00:00.182) 0:00:30.236 ***** 13118 1727204092.93701: entering _queue_task() for managed-node2/package 13118 1727204092.94053: worker is 1 (out of 1 available) 13118 1727204092.94347: exiting _queue_task() for managed-node2/package 13118 1727204092.94361: done queuing things up, now waiting for results queue to drain 13118 1727204092.94362: waiting for pending results... 13118 1727204092.95022: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Install packages 13118 1727204092.95311: in run() - task 0affcd87-79f5-56a3-0a64-000000000084 13118 1727204092.95389: variable 'ansible_search_path' from source: unknown 13118 1727204092.95398: variable 'ansible_search_path' from source: unknown 13118 1727204092.95441: calling self._execute() 13118 1727204092.95696: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204092.95822: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204092.95838: variable 'omit' from source: magic vars 13118 1727204092.96467: variable 'ansible_distribution_major_version' from source: facts 13118 1727204092.96598: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204092.97004: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204092.97924: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204092.98105: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204092.98145: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204092.98235: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204092.98552: variable 'network_packages' from source: role '' defaults 13118 1727204092.98787: variable '__network_provider_setup' from source: role '' defaults 13118 1727204092.98804: variable '__network_service_name_default_nm' from source: role '' defaults 13118 1727204092.98990: variable '__network_service_name_default_nm' from source: role '' defaults 13118 1727204092.99004: variable '__network_packages_default_nm' from source: role '' defaults 13118 1727204092.99110: variable '__network_packages_default_nm' from source: role '' defaults 13118 1727204092.99662: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204093.04444: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204093.04734: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204093.04786: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204093.04826: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204093.04860: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204093.05084: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.05274: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.05310: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.05410: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.05497: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.05567: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.05739: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.05772: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.05818: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.05951: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.06323: variable '__network_packages_default_gobject_packages' from source: role '' defaults 13118 1727204093.06623: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.06726: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.06757: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.06850: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.06890: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.07017: variable 'ansible_python' from source: facts 13118 1727204093.07161: variable '__network_packages_default_wpa_supplicant' from source: role '' defaults 13118 1727204093.07424: variable '__network_wpa_supplicant_required' from source: role '' defaults 13118 1727204093.07767: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13118 1727204093.08097: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.08236: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.08268: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.08308: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.08322: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.08405: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.08431: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.08570: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.08611: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.08625: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.09009: variable 'network_connections' from source: task vars 13118 1727204093.09016: variable 'port2_profile' from source: play vars 13118 1727204093.09346: variable 'port2_profile' from source: play vars 13118 1727204093.09358: variable 'port1_profile' from source: play vars 13118 1727204093.09671: variable 'port1_profile' from source: play vars 13118 1727204093.09681: variable 'controller_profile' from source: play vars 13118 1727204093.09896: variable 'controller_profile' from source: play vars 13118 1727204093.09971: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204093.10115: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204093.10146: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.10179: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204093.10343: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204093.10927: variable 'network_connections' from source: task vars 13118 1727204093.10935: variable 'port2_profile' from source: play vars 13118 1727204093.11146: variable 'port2_profile' from source: play vars 13118 1727204093.11156: variable 'port1_profile' from source: play vars 13118 1727204093.11369: variable 'port1_profile' from source: play vars 13118 1727204093.11379: variable 'controller_profile' from source: play vars 13118 1727204093.11654: variable 'controller_profile' from source: play vars 13118 1727204093.11684: variable '__network_packages_default_wireless' from source: role '' defaults 13118 1727204093.11981: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204093.12626: variable 'network_connections' from source: task vars 13118 1727204093.12633: variable 'port2_profile' from source: play vars 13118 1727204093.12809: variable 'port2_profile' from source: play vars 13118 1727204093.12817: variable 'port1_profile' from source: play vars 13118 1727204093.12881: variable 'port1_profile' from source: play vars 13118 1727204093.12889: variable 'controller_profile' from source: play vars 13118 1727204093.13070: variable 'controller_profile' from source: play vars 13118 1727204093.13094: variable '__network_packages_default_team' from source: role '' defaults 13118 1727204093.13289: variable '__network_team_connections_defined' from source: role '' defaults 13118 1727204093.13936: variable 'network_connections' from source: task vars 13118 1727204093.13941: variable 'port2_profile' from source: play vars 13118 1727204093.14118: variable 'port2_profile' from source: play vars 13118 1727204093.14125: variable 'port1_profile' from source: play vars 13118 1727204093.14187: variable 'port1_profile' from source: play vars 13118 1727204093.14195: variable 'controller_profile' from source: play vars 13118 1727204093.14376: variable 'controller_profile' from source: play vars 13118 1727204093.14548: variable '__network_service_name_default_initscripts' from source: role '' defaults 13118 1727204093.14611: variable '__network_service_name_default_initscripts' from source: role '' defaults 13118 1727204093.14617: variable '__network_packages_default_initscripts' from source: role '' defaults 13118 1727204093.14792: variable '__network_packages_default_initscripts' from source: role '' defaults 13118 1727204093.15287: variable '__network_packages_default_initscripts_bridge' from source: role '' defaults 13118 1727204093.16357: variable 'network_connections' from source: task vars 13118 1727204093.16362: variable 'port2_profile' from source: play vars 13118 1727204093.16536: variable 'port2_profile' from source: play vars 13118 1727204093.16542: variable 'port1_profile' from source: play vars 13118 1727204093.16712: variable 'port1_profile' from source: play vars 13118 1727204093.16719: variable 'controller_profile' from source: play vars 13118 1727204093.16780: variable 'controller_profile' from source: play vars 13118 1727204093.16788: variable 'ansible_distribution' from source: facts 13118 1727204093.16791: variable '__network_rh_distros' from source: role '' defaults 13118 1727204093.16795: variable 'ansible_distribution_major_version' from source: facts 13118 1727204093.16926: variable '__network_packages_default_initscripts_network_scripts' from source: role '' defaults 13118 1727204093.17215: variable 'ansible_distribution' from source: facts 13118 1727204093.17220: variable '__network_rh_distros' from source: role '' defaults 13118 1727204093.17223: variable 'ansible_distribution_major_version' from source: facts 13118 1727204093.17237: variable '__network_packages_default_initscripts_dhcp_client' from source: role '' defaults 13118 1727204093.17955: variable 'ansible_distribution' from source: facts 13118 1727204093.17958: variable '__network_rh_distros' from source: role '' defaults 13118 1727204093.17961: variable 'ansible_distribution_major_version' from source: facts 13118 1727204093.17963: variable 'network_provider' from source: set_fact 13118 1727204093.17967: variable 'ansible_facts' from source: unknown 13118 1727204093.19578: Evaluated conditional (not network_packages is subset(ansible_facts.packages.keys())): False 13118 1727204093.19582: when evaluation is False, skipping this task 13118 1727204093.19585: _execute() done 13118 1727204093.19587: dumping result to json 13118 1727204093.19589: done dumping result, returning 13118 1727204093.19598: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Install packages [0affcd87-79f5-56a3-0a64-000000000084] 13118 1727204093.19604: sending task result for task 0affcd87-79f5-56a3-0a64-000000000084 skipping: [managed-node2] => { "changed": false, "false_condition": "not network_packages is subset(ansible_facts.packages.keys())", "skip_reason": "Conditional result was False" } 13118 1727204093.19773: no more pending results, returning what we have 13118 1727204093.19778: results queue empty 13118 1727204093.19779: checking for any_errors_fatal 13118 1727204093.19786: done checking for any_errors_fatal 13118 1727204093.19787: checking for max_fail_percentage 13118 1727204093.19788: done checking for max_fail_percentage 13118 1727204093.19789: checking to see if all hosts have failed and the running result is not ok 13118 1727204093.19790: done checking to see if all hosts have failed 13118 1727204093.19790: getting the remaining hosts for this loop 13118 1727204093.19792: done getting the remaining hosts for this loop 13118 1727204093.19795: getting the next task for host managed-node2 13118 1727204093.19802: done getting next task for host managed-node2 13118 1727204093.19810: ^ task is: TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable 13118 1727204093.19814: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204093.19832: getting variables 13118 1727204093.19833: in VariableManager get_vars() 13118 1727204093.19875: Calling all_inventory to load vars for managed-node2 13118 1727204093.19879: Calling groups_inventory to load vars for managed-node2 13118 1727204093.19881: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204093.19892: Calling all_plugins_play to load vars for managed-node2 13118 1727204093.19895: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204093.19898: Calling groups_plugins_play to load vars for managed-node2 13118 1727204093.20672: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000084 13118 1727204093.20675: WORKER PROCESS EXITING 13118 1727204093.22173: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204093.25810: done with get_vars() 13118 1727204093.25846: done getting variables 13118 1727204093.26182: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:85 Tuesday 24 September 2024 14:54:53 -0400 (0:00:00.325) 0:00:30.562 ***** 13118 1727204093.26222: entering _queue_task() for managed-node2/package 13118 1727204093.27319: worker is 1 (out of 1 available) 13118 1727204093.27331: exiting _queue_task() for managed-node2/package 13118 1727204093.27342: done queuing things up, now waiting for results queue to drain 13118 1727204093.27344: waiting for pending results... 13118 1727204093.27428: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable 13118 1727204093.27804: in run() - task 0affcd87-79f5-56a3-0a64-000000000085 13118 1727204093.27817: variable 'ansible_search_path' from source: unknown 13118 1727204093.27821: variable 'ansible_search_path' from source: unknown 13118 1727204093.27862: calling self._execute() 13118 1727204093.28078: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204093.28083: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204093.28209: variable 'omit' from source: magic vars 13118 1727204093.28943: variable 'ansible_distribution_major_version' from source: facts 13118 1727204093.28955: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204093.29308: variable 'network_state' from source: role '' defaults 13118 1727204093.29320: Evaluated conditional (network_state != {}): False 13118 1727204093.29323: when evaluation is False, skipping this task 13118 1727204093.29325: _execute() done 13118 1727204093.29328: dumping result to json 13118 1727204093.29335: done dumping result, returning 13118 1727204093.29343: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable [0affcd87-79f5-56a3-0a64-000000000085] 13118 1727204093.29350: sending task result for task 0affcd87-79f5-56a3-0a64-000000000085 skipping: [managed-node2] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13118 1727204093.29521: no more pending results, returning what we have 13118 1727204093.29525: results queue empty 13118 1727204093.29526: checking for any_errors_fatal 13118 1727204093.29534: done checking for any_errors_fatal 13118 1727204093.29535: checking for max_fail_percentage 13118 1727204093.29537: done checking for max_fail_percentage 13118 1727204093.29538: checking to see if all hosts have failed and the running result is not ok 13118 1727204093.29539: done checking to see if all hosts have failed 13118 1727204093.29539: getting the remaining hosts for this loop 13118 1727204093.29541: done getting the remaining hosts for this loop 13118 1727204093.29544: getting the next task for host managed-node2 13118 1727204093.29553: done getting next task for host managed-node2 13118 1727204093.29557: ^ task is: TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable 13118 1727204093.29562: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=14, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204093.29583: getting variables 13118 1727204093.29585: in VariableManager get_vars() 13118 1727204093.29630: Calling all_inventory to load vars for managed-node2 13118 1727204093.29633: Calling groups_inventory to load vars for managed-node2 13118 1727204093.29636: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204093.29649: Calling all_plugins_play to load vars for managed-node2 13118 1727204093.29652: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204093.29656: Calling groups_plugins_play to load vars for managed-node2 13118 1727204093.30473: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000085 13118 1727204093.30477: WORKER PROCESS EXITING 13118 1727204093.32409: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204093.35636: done with get_vars() 13118 1727204093.35974: done getting variables 13118 1727204093.36037: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:96 Tuesday 24 September 2024 14:54:53 -0400 (0:00:00.098) 0:00:30.660 ***** 13118 1727204093.36077: entering _queue_task() for managed-node2/package 13118 1727204093.36821: worker is 1 (out of 1 available) 13118 1727204093.36835: exiting _queue_task() for managed-node2/package 13118 1727204093.36848: done queuing things up, now waiting for results queue to drain 13118 1727204093.36849: waiting for pending results... 13118 1727204093.37307: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable 13118 1727204093.37671: in run() - task 0affcd87-79f5-56a3-0a64-000000000086 13118 1727204093.37683: variable 'ansible_search_path' from source: unknown 13118 1727204093.37686: variable 'ansible_search_path' from source: unknown 13118 1727204093.37721: calling self._execute() 13118 1727204093.37931: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204093.37939: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204093.37949: variable 'omit' from source: magic vars 13118 1727204093.38769: variable 'ansible_distribution_major_version' from source: facts 13118 1727204093.38781: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204093.39014: variable 'network_state' from source: role '' defaults 13118 1727204093.39024: Evaluated conditional (network_state != {}): False 13118 1727204093.39027: when evaluation is False, skipping this task 13118 1727204093.39030: _execute() done 13118 1727204093.39032: dumping result to json 13118 1727204093.39038: done dumping result, returning 13118 1727204093.39046: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable [0affcd87-79f5-56a3-0a64-000000000086] 13118 1727204093.39169: sending task result for task 0affcd87-79f5-56a3-0a64-000000000086 13118 1727204093.39268: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000086 13118 1727204093.39272: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13118 1727204093.39318: no more pending results, returning what we have 13118 1727204093.39324: results queue empty 13118 1727204093.39325: checking for any_errors_fatal 13118 1727204093.39332: done checking for any_errors_fatal 13118 1727204093.39332: checking for max_fail_percentage 13118 1727204093.39334: done checking for max_fail_percentage 13118 1727204093.39335: checking to see if all hosts have failed and the running result is not ok 13118 1727204093.39336: done checking to see if all hosts have failed 13118 1727204093.39337: getting the remaining hosts for this loop 13118 1727204093.39339: done getting the remaining hosts for this loop 13118 1727204093.39342: getting the next task for host managed-node2 13118 1727204093.39351: done getting next task for host managed-node2 13118 1727204093.39356: ^ task is: TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces 13118 1727204093.39362: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204093.39385: getting variables 13118 1727204093.39387: in VariableManager get_vars() 13118 1727204093.39426: Calling all_inventory to load vars for managed-node2 13118 1727204093.39428: Calling groups_inventory to load vars for managed-node2 13118 1727204093.39430: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204093.39441: Calling all_plugins_play to load vars for managed-node2 13118 1727204093.39444: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204093.39446: Calling groups_plugins_play to load vars for managed-node2 13118 1727204093.41922: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204093.45566: done with get_vars() 13118 1727204093.45592: done getting variables 13118 1727204093.45655: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:109 Tuesday 24 September 2024 14:54:53 -0400 (0:00:00.098) 0:00:30.758 ***** 13118 1727204093.45901: entering _queue_task() for managed-node2/service 13118 1727204093.46645: worker is 1 (out of 1 available) 13118 1727204093.46657: exiting _queue_task() for managed-node2/service 13118 1727204093.46672: done queuing things up, now waiting for results queue to drain 13118 1727204093.46673: waiting for pending results... 13118 1727204093.48706: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces 13118 1727204093.49079: in run() - task 0affcd87-79f5-56a3-0a64-000000000087 13118 1727204093.49092: variable 'ansible_search_path' from source: unknown 13118 1727204093.49096: variable 'ansible_search_path' from source: unknown 13118 1727204093.49131: calling self._execute() 13118 1727204093.49344: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204093.49348: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204093.49360: variable 'omit' from source: magic vars 13118 1727204093.50219: variable 'ansible_distribution_major_version' from source: facts 13118 1727204093.50232: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204093.50468: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204093.50896: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204093.56294: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204093.56478: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204093.56518: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204093.56556: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204093.56585: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204093.56779: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.56921: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.56951: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.56994: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.57009: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.57173: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.57196: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.57222: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.57380: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.57395: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.57437: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.57552: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.57742: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.57745: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.57768: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.58099: variable 'network_connections' from source: task vars 13118 1727204093.58227: variable 'port2_profile' from source: play vars 13118 1727204093.58299: variable 'port2_profile' from source: play vars 13118 1727204093.58310: variable 'port1_profile' from source: play vars 13118 1727204093.58489: variable 'port1_profile' from source: play vars 13118 1727204093.58498: variable 'controller_profile' from source: play vars 13118 1727204093.58675: variable 'controller_profile' from source: play vars 13118 1727204093.58744: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204093.59159: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204093.59329: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204093.59361: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204093.59391: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204093.59551: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204093.59576: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204093.59603: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.59745: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204093.59799: variable '__network_team_connections_defined' from source: role '' defaults 13118 1727204093.60397: variable 'network_connections' from source: task vars 13118 1727204093.60401: variable 'port2_profile' from source: play vars 13118 1727204093.60470: variable 'port2_profile' from source: play vars 13118 1727204093.60478: variable 'port1_profile' from source: play vars 13118 1727204093.60651: variable 'port1_profile' from source: play vars 13118 1727204093.60659: variable 'controller_profile' from source: play vars 13118 1727204093.60836: variable 'controller_profile' from source: play vars 13118 1727204093.60862: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13118 1727204093.60876: when evaluation is False, skipping this task 13118 1727204093.60879: _execute() done 13118 1727204093.60881: dumping result to json 13118 1727204093.60884: done dumping result, returning 13118 1727204093.60886: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces [0affcd87-79f5-56a3-0a64-000000000087] 13118 1727204093.60888: sending task result for task 0affcd87-79f5-56a3-0a64-000000000087 13118 1727204093.61100: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000087 13118 1727204093.61103: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13118 1727204093.61180: no more pending results, returning what we have 13118 1727204093.61185: results queue empty 13118 1727204093.61186: checking for any_errors_fatal 13118 1727204093.61193: done checking for any_errors_fatal 13118 1727204093.61194: checking for max_fail_percentage 13118 1727204093.61196: done checking for max_fail_percentage 13118 1727204093.61197: checking to see if all hosts have failed and the running result is not ok 13118 1727204093.61198: done checking to see if all hosts have failed 13118 1727204093.61199: getting the remaining hosts for this loop 13118 1727204093.61200: done getting the remaining hosts for this loop 13118 1727204093.61205: getting the next task for host managed-node2 13118 1727204093.61216: done getting next task for host managed-node2 13118 1727204093.61221: ^ task is: TASK: fedora.linux_system_roles.network : Enable and start NetworkManager 13118 1727204093.61225: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204093.61244: getting variables 13118 1727204093.61246: in VariableManager get_vars() 13118 1727204093.61287: Calling all_inventory to load vars for managed-node2 13118 1727204093.61290: Calling groups_inventory to load vars for managed-node2 13118 1727204093.61292: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204093.61302: Calling all_plugins_play to load vars for managed-node2 13118 1727204093.61304: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204093.61307: Calling groups_plugins_play to load vars for managed-node2 13118 1727204093.64877: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204093.68075: done with get_vars() 13118 1727204093.68108: done getting variables 13118 1727204093.68288: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable and start NetworkManager] ***** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:122 Tuesday 24 September 2024 14:54:53 -0400 (0:00:00.224) 0:00:30.983 ***** 13118 1727204093.68324: entering _queue_task() for managed-node2/service 13118 1727204093.69160: worker is 1 (out of 1 available) 13118 1727204093.69173: exiting _queue_task() for managed-node2/service 13118 1727204093.69185: done queuing things up, now waiting for results queue to drain 13118 1727204093.69186: waiting for pending results... 13118 1727204093.70178: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Enable and start NetworkManager 13118 1727204093.70551: in run() - task 0affcd87-79f5-56a3-0a64-000000000088 13118 1727204093.70566: variable 'ansible_search_path' from source: unknown 13118 1727204093.70572: variable 'ansible_search_path' from source: unknown 13118 1727204093.70608: calling self._execute() 13118 1727204093.70827: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204093.70836: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204093.70968: variable 'omit' from source: magic vars 13118 1727204093.71848: variable 'ansible_distribution_major_version' from source: facts 13118 1727204093.71860: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204093.72268: variable 'network_provider' from source: set_fact 13118 1727204093.72278: variable 'network_state' from source: role '' defaults 13118 1727204093.72290: Evaluated conditional (network_provider == "nm" or network_state != {}): True 13118 1727204093.72297: variable 'omit' from source: magic vars 13118 1727204093.72492: variable 'omit' from source: magic vars 13118 1727204093.72525: variable 'network_service_name' from source: role '' defaults 13118 1727204093.72717: variable 'network_service_name' from source: role '' defaults 13118 1727204093.72953: variable '__network_provider_setup' from source: role '' defaults 13118 1727204093.72959: variable '__network_service_name_default_nm' from source: role '' defaults 13118 1727204093.73117: variable '__network_service_name_default_nm' from source: role '' defaults 13118 1727204093.73135: variable '__network_packages_default_nm' from source: role '' defaults 13118 1727204093.73376: variable '__network_packages_default_nm' from source: role '' defaults 13118 1727204093.73950: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204093.78989: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204093.79192: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204093.79348: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204093.79388: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204093.79414: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204093.79620: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.79650: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.79797: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.79843: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.79857: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.80022: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.80048: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.80073: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.80229: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.80248: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.80733: variable '__network_packages_default_gobject_packages' from source: role '' defaults 13118 1727204093.81096: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.81122: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.81149: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.81303: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.81324: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.81531: variable 'ansible_python' from source: facts 13118 1727204093.81557: variable '__network_packages_default_wpa_supplicant' from source: role '' defaults 13118 1727204093.81653: variable '__network_wpa_supplicant_required' from source: role '' defaults 13118 1727204093.81742: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13118 1727204093.81883: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.81906: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.81929: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.81981: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.81995: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.82538: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204093.82560: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204093.82586: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.82746: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204093.82760: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204093.83027: variable 'network_connections' from source: task vars 13118 1727204093.83062: variable 'port2_profile' from source: play vars 13118 1727204093.83143: variable 'port2_profile' from source: play vars 13118 1727204093.83159: variable 'port1_profile' from source: play vars 13118 1727204093.83232: variable 'port1_profile' from source: play vars 13118 1727204093.83254: variable 'controller_profile' from source: play vars 13118 1727204093.83332: variable 'controller_profile' from source: play vars 13118 1727204093.83450: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204093.84066: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204093.84124: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204093.84178: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204093.84218: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204093.84294: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204093.84322: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204093.84369: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204093.84402: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204093.84467: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204093.84789: variable 'network_connections' from source: task vars 13118 1727204093.84796: variable 'port2_profile' from source: play vars 13118 1727204093.84877: variable 'port2_profile' from source: play vars 13118 1727204093.84894: variable 'port1_profile' from source: play vars 13118 1727204093.84970: variable 'port1_profile' from source: play vars 13118 1727204093.84988: variable 'controller_profile' from source: play vars 13118 1727204093.85068: variable 'controller_profile' from source: play vars 13118 1727204093.85110: variable '__network_packages_default_wireless' from source: role '' defaults 13118 1727204093.85196: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204093.85535: variable 'network_connections' from source: task vars 13118 1727204093.85540: variable 'port2_profile' from source: play vars 13118 1727204093.85616: variable 'port2_profile' from source: play vars 13118 1727204093.85624: variable 'port1_profile' from source: play vars 13118 1727204093.85710: variable 'port1_profile' from source: play vars 13118 1727204093.85717: variable 'controller_profile' from source: play vars 13118 1727204093.85796: variable 'controller_profile' from source: play vars 13118 1727204093.85821: variable '__network_packages_default_team' from source: role '' defaults 13118 1727204093.85912: variable '__network_team_connections_defined' from source: role '' defaults 13118 1727204093.86242: variable 'network_connections' from source: task vars 13118 1727204093.86245: variable 'port2_profile' from source: play vars 13118 1727204093.86325: variable 'port2_profile' from source: play vars 13118 1727204093.86332: variable 'port1_profile' from source: play vars 13118 1727204093.86407: variable 'port1_profile' from source: play vars 13118 1727204093.86415: variable 'controller_profile' from source: play vars 13118 1727204093.86491: variable 'controller_profile' from source: play vars 13118 1727204093.86559: variable '__network_service_name_default_initscripts' from source: role '' defaults 13118 1727204093.86623: variable '__network_service_name_default_initscripts' from source: role '' defaults 13118 1727204093.86629: variable '__network_packages_default_initscripts' from source: role '' defaults 13118 1727204093.86697: variable '__network_packages_default_initscripts' from source: role '' defaults 13118 1727204093.86944: variable '__network_packages_default_initscripts_bridge' from source: role '' defaults 13118 1727204093.87868: variable 'network_connections' from source: task vars 13118 1727204093.87876: variable 'port2_profile' from source: play vars 13118 1727204093.88057: variable 'port2_profile' from source: play vars 13118 1727204093.88065: variable 'port1_profile' from source: play vars 13118 1727204093.88238: variable 'port1_profile' from source: play vars 13118 1727204093.88246: variable 'controller_profile' from source: play vars 13118 1727204093.88312: variable 'controller_profile' from source: play vars 13118 1727204093.88320: variable 'ansible_distribution' from source: facts 13118 1727204093.88323: variable '__network_rh_distros' from source: role '' defaults 13118 1727204093.88329: variable 'ansible_distribution_major_version' from source: facts 13118 1727204093.88467: variable '__network_packages_default_initscripts_network_scripts' from source: role '' defaults 13118 1727204093.88870: variable 'ansible_distribution' from source: facts 13118 1727204093.88873: variable '__network_rh_distros' from source: role '' defaults 13118 1727204093.88876: variable 'ansible_distribution_major_version' from source: facts 13118 1727204093.88967: variable '__network_packages_default_initscripts_dhcp_client' from source: role '' defaults 13118 1727204093.89150: variable 'ansible_distribution' from source: facts 13118 1727204093.89154: variable '__network_rh_distros' from source: role '' defaults 13118 1727204093.89161: variable 'ansible_distribution_major_version' from source: facts 13118 1727204093.89199: variable 'network_provider' from source: set_fact 13118 1727204093.89227: variable 'omit' from source: magic vars 13118 1727204093.89277: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204093.89303: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204093.89352: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204093.89384: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204093.89388: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204093.89415: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204093.89420: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204093.89425: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204093.89544: Set connection var ansible_timeout to 10 13118 1727204093.89556: Set connection var ansible_pipelining to False 13118 1727204093.89558: Set connection var ansible_connection to ssh 13118 1727204093.89567: Set connection var ansible_shell_executable to /bin/sh 13118 1727204093.89573: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204093.89581: Set connection var ansible_shell_type to sh 13118 1727204093.89606: variable 'ansible_shell_executable' from source: unknown 13118 1727204093.89610: variable 'ansible_connection' from source: unknown 13118 1727204093.89613: variable 'ansible_module_compression' from source: unknown 13118 1727204093.89615: variable 'ansible_shell_type' from source: unknown 13118 1727204093.89617: variable 'ansible_shell_executable' from source: unknown 13118 1727204093.89619: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204093.89624: variable 'ansible_pipelining' from source: unknown 13118 1727204093.89626: variable 'ansible_timeout' from source: unknown 13118 1727204093.89631: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204093.89754: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204093.89768: variable 'omit' from source: magic vars 13118 1727204093.89774: starting attempt loop 13118 1727204093.89780: running the handler 13118 1727204093.89875: variable 'ansible_facts' from source: unknown 13118 1727204093.90731: _low_level_execute_command(): starting 13118 1727204093.90748: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204093.92565: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204093.92589: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204093.92601: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204093.92615: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204093.92658: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204093.92666: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204093.92676: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204093.92690: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204093.92697: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204093.92705: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204093.92714: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204093.92727: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204093.92741: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204093.92748: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204093.92757: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204093.92771: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204093.92848: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204093.92862: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204093.92867: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204093.92945: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204093.94599: stdout chunk (state=3): >>>/root <<< 13118 1727204093.94902: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204093.94974: stderr chunk (state=3): >>><<< 13118 1727204093.94979: stdout chunk (state=3): >>><<< 13118 1727204093.95000: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204093.95013: _low_level_execute_command(): starting 13118 1727204093.95019: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204093.950003-15665-9280954286544 `" && echo ansible-tmp-1727204093.950003-15665-9280954286544="` echo /root/.ansible/tmp/ansible-tmp-1727204093.950003-15665-9280954286544 `" ) && sleep 0' 13118 1727204093.96347: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204093.96982: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204093.96993: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204093.97008: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204093.97056: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204093.97059: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204093.97071: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204093.97085: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204093.97093: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204093.97100: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204093.97107: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204093.97116: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204093.97128: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204093.97138: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204093.97145: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204093.97155: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204093.97232: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204093.97249: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204093.97259: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204093.97379: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204093.99194: stdout chunk (state=3): >>>ansible-tmp-1727204093.950003-15665-9280954286544=/root/.ansible/tmp/ansible-tmp-1727204093.950003-15665-9280954286544 <<< 13118 1727204093.99397: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204093.99401: stdout chunk (state=3): >>><<< 13118 1727204093.99408: stderr chunk (state=3): >>><<< 13118 1727204093.99426: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204093.950003-15665-9280954286544=/root/.ansible/tmp/ansible-tmp-1727204093.950003-15665-9280954286544 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204093.99468: variable 'ansible_module_compression' from source: unknown 13118 1727204093.99523: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.systemd-ZIP_DEFLATED 13118 1727204093.99588: variable 'ansible_facts' from source: unknown 13118 1727204093.99784: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204093.950003-15665-9280954286544/AnsiballZ_systemd.py 13118 1727204094.00661: Sending initial data 13118 1727204094.00668: Sent initial data (153 bytes) 13118 1727204094.03328: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.03337: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.03484: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204094.03491: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.03556: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.03639: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204094.03768: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204094.03775: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204094.03851: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204094.05584: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 <<< 13118 1727204094.05588: stderr chunk (state=3): >>>debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204094.05629: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204094.05667: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmps4ok2ism /root/.ansible/tmp/ansible-tmp-1727204093.950003-15665-9280954286544/AnsiballZ_systemd.py <<< 13118 1727204094.05702: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204094.08578: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204094.08767: stderr chunk (state=3): >>><<< 13118 1727204094.08771: stdout chunk (state=3): >>><<< 13118 1727204094.08793: done transferring module to remote 13118 1727204094.08821: _low_level_execute_command(): starting 13118 1727204094.08824: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204093.950003-15665-9280954286544/ /root/.ansible/tmp/ansible-tmp-1727204093.950003-15665-9280954286544/AnsiballZ_systemd.py && sleep 0' 13118 1727204094.10386: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204094.10538: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.11133: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.11146: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.11192: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204094.11589: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204094.11599: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.11619: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204094.11627: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204094.11634: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204094.11642: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.11652: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.11665: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.11675: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204094.11682: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204094.11692: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.12198: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204094.12213: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204094.12222: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204094.12410: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204094.14188: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204094.14249: stderr chunk (state=3): >>><<< 13118 1727204094.14257: stdout chunk (state=3): >>><<< 13118 1727204094.14272: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204094.14275: _low_level_execute_command(): starting 13118 1727204094.14281: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204093.950003-15665-9280954286544/AnsiballZ_systemd.py && sleep 0' 13118 1727204094.15987: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204094.15999: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.16011: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.16025: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.16063: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204094.16072: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204094.16082: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.16096: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204094.16106: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204094.16112: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204094.16120: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.16129: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.16140: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.16147: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204094.16153: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204094.16162: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.16239: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204094.16253: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204094.16266: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204094.16353: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204094.41952: stdout chunk (state=3): >>> {"name": "NetworkManager", "changed": false, "status": {"Type": "dbus", "ExitType": "main", "Restart": "on-failure", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "10min", "TimeoutStopUSec": "1min 30s", "TimeoutAbortUSec": "1min 30s", "TimeoutStartFailureMode": "terminate", "TimeoutStopFailureMode": "terminate", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "WatchdogUSec": "0", "WatchdogTimestampMonotonic": "0", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "6823", "ControlPID": "0", "BusName": "org.freedesktop.NetworkManager", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "ReloadResult": "success", "CleanResult": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "OOMPolicy": "stop", "ReloadSignal": "1", "ExecMainStartTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ExecMainStartTimestampMonotonic": "319366198", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "6823", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReload": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/NetworkManager.service", "ControlGroupId": "3602", "MemoryCurrent": "6705152", "MemoryAvailable": "infinity", "CPUUsageNSec": "565577000", "TasksCurrent": "3", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "Delegate": "no", "CPUAccounting": "yes", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "MemoryHigh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22342", "IPAccounting": "no", "ManagedOOMSwap": "auto", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "65536", "LimitNOFILESoft": "65536", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "13964", "LimitNPROCSoft": "13964", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "13964", "LimitSIGPENDINGSoft": "13964", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "<<< 13118 1727204094.41961: stdout chunk (state=3): >>>0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "CoredumpFilter": "0x33", "Nice": "0", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_dac_override cap_kill cap_setgid cap_setuid cap_net_bind_service cap_net_admin cap_net_raw cap_sys_module cap_sys_chroot cap_audit_write", "DynamicUser": "no", "RemoveIPC": "no", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectClock": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectKernelLogs": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "PrivateIPC": "no", "ProtectHome": "read-only", "ProtectSystem": "yes", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "2147483646", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "TimeoutCleanUSec": "infinity", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "ProtectProc": "default", "ProcSubset": "all", "ProtectHostname": "no", "KillMode": "process", "KillSignal": "15", "RestartKillSignal": "15", "FinalKillSignal": "9", "SendSIGKILL": "yes", "SendSIGHUP": "no", "WatchdogSignal": "6", "Id": "NetworkManager.service", "Names": "NetworkManager.service", "Requires": "dbus.socket system.slice sysinit.target", "Wants": "network.target", "BindsTo": "dbus-broker.service", "RequiredBy": "NetworkManager-wait-online.service", "WantedBy": "multi-user.target", "Conflicts": "shutdown.target", "Before": "NetworkManager-wait-online.service cloud-init.service network.target network.service multi-user.target shutdown.target", "After": "systemd-journald.socket network-pre.target dbus-broker.service cloud-init-local.service system.slice basic.target dbus.socket sysinit.target", "Documentation": "\"man:NetworkManager(8)\"", "Description": "Network Manager", "AccessSELinuxContext": "system_u:object_r:NetworkManager_unit_file_t:s0", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/NetworkManager.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Tue 2024-09-24 14:54:30 EDT", "StateChangeTimestampMonotonic": "433536261", "InactiveExitTimestamp": "Tue 2024-09-24 14:52:36 EDT", "InactiveExitTimestampMonotonic": "319366492", "ActiveEnterTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ActiveEnterTimestampMonotonic": "319444795", "ActiveExitTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ActiveExitTimestampMonotonic": "319337881", "InactiveEnterTimestamp": "Tue 2024-09-24 14:52:36 EDT", "InactiveEnterTimestampMonotonic": "319361759", "CanStart": "yes", "CanStop": "yes", "CanReload": "yes", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnSuccessJobMode": "fail", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ConditionTimestampMonotonic": "319362324", "AssertTimestamp": "Tue 2024-09-24 14:52:36 EDT", "AssertTimestampMonotonic": "319362327", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "bc82db972fb14d0fb9ce19d409aedafe", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "NetworkManager", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} <<< 13118 1727204094.43537: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204094.43542: stdout chunk (state=3): >>><<< 13118 1727204094.43544: stderr chunk (state=3): >>><<< 13118 1727204094.43557: _low_level_execute_command() done: rc=0, stdout= {"name": "NetworkManager", "changed": false, "status": {"Type": "dbus", "ExitType": "main", "Restart": "on-failure", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "10min", "TimeoutStopUSec": "1min 30s", "TimeoutAbortUSec": "1min 30s", "TimeoutStartFailureMode": "terminate", "TimeoutStopFailureMode": "terminate", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "WatchdogUSec": "0", "WatchdogTimestampMonotonic": "0", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "6823", "ControlPID": "0", "BusName": "org.freedesktop.NetworkManager", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "ReloadResult": "success", "CleanResult": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "OOMPolicy": "stop", "ReloadSignal": "1", "ExecMainStartTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ExecMainStartTimestampMonotonic": "319366198", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "6823", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReload": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/NetworkManager.service", "ControlGroupId": "3602", "MemoryCurrent": "6705152", "MemoryAvailable": "infinity", "CPUUsageNSec": "565577000", "TasksCurrent": "3", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "Delegate": "no", "CPUAccounting": "yes", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "MemoryHigh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22342", "IPAccounting": "no", "ManagedOOMSwap": "auto", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "65536", "LimitNOFILESoft": "65536", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "13964", "LimitNPROCSoft": "13964", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "13964", "LimitSIGPENDINGSoft": "13964", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "CoredumpFilter": "0x33", "Nice": "0", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_dac_override cap_kill cap_setgid cap_setuid cap_net_bind_service cap_net_admin cap_net_raw cap_sys_module cap_sys_chroot cap_audit_write", "DynamicUser": "no", "RemoveIPC": "no", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectClock": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectKernelLogs": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "PrivateIPC": "no", "ProtectHome": "read-only", "ProtectSystem": "yes", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "2147483646", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "TimeoutCleanUSec": "infinity", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "ProtectProc": "default", "ProcSubset": "all", "ProtectHostname": "no", "KillMode": "process", "KillSignal": "15", "RestartKillSignal": "15", "FinalKillSignal": "9", "SendSIGKILL": "yes", "SendSIGHUP": "no", "WatchdogSignal": "6", "Id": "NetworkManager.service", "Names": "NetworkManager.service", "Requires": "dbus.socket system.slice sysinit.target", "Wants": "network.target", "BindsTo": "dbus-broker.service", "RequiredBy": "NetworkManager-wait-online.service", "WantedBy": "multi-user.target", "Conflicts": "shutdown.target", "Before": "NetworkManager-wait-online.service cloud-init.service network.target network.service multi-user.target shutdown.target", "After": "systemd-journald.socket network-pre.target dbus-broker.service cloud-init-local.service system.slice basic.target dbus.socket sysinit.target", "Documentation": "\"man:NetworkManager(8)\"", "Description": "Network Manager", "AccessSELinuxContext": "system_u:object_r:NetworkManager_unit_file_t:s0", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/NetworkManager.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Tue 2024-09-24 14:54:30 EDT", "StateChangeTimestampMonotonic": "433536261", "InactiveExitTimestamp": "Tue 2024-09-24 14:52:36 EDT", "InactiveExitTimestampMonotonic": "319366492", "ActiveEnterTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ActiveEnterTimestampMonotonic": "319444795", "ActiveExitTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ActiveExitTimestampMonotonic": "319337881", "InactiveEnterTimestamp": "Tue 2024-09-24 14:52:36 EDT", "InactiveEnterTimestampMonotonic": "319361759", "CanStart": "yes", "CanStop": "yes", "CanReload": "yes", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnSuccessJobMode": "fail", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Tue 2024-09-24 14:52:36 EDT", "ConditionTimestampMonotonic": "319362324", "AssertTimestamp": "Tue 2024-09-24 14:52:36 EDT", "AssertTimestampMonotonic": "319362327", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "bc82db972fb14d0fb9ce19d409aedafe", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "NetworkManager", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204094.43737: done with _execute_module (ansible.legacy.systemd, {'name': 'NetworkManager', 'state': 'started', 'enabled': True, '_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.systemd', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204093.950003-15665-9280954286544/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204094.43754: _low_level_execute_command(): starting 13118 1727204094.43758: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204093.950003-15665-9280954286544/ > /dev/null 2>&1 && sleep 0' 13118 1727204094.44595: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204094.44604: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.44613: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.44626: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.44662: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204094.44671: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204094.44682: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.44695: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204094.44702: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204094.44709: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204094.44733: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.44736: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.44738: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.44750: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204094.44753: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204094.44785: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.44841: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204094.44848: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204094.44855: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204094.44928: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204094.46769: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204094.46774: stderr chunk (state=3): >>><<< 13118 1727204094.46776: stdout chunk (state=3): >>><<< 13118 1727204094.46797: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204094.46804: handler run complete 13118 1727204094.46871: attempt loop complete, returning result 13118 1727204094.46875: _execute() done 13118 1727204094.46878: dumping result to json 13118 1727204094.46896: done dumping result, returning 13118 1727204094.46907: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Enable and start NetworkManager [0affcd87-79f5-56a3-0a64-000000000088] 13118 1727204094.46912: sending task result for task 0affcd87-79f5-56a3-0a64-000000000088 13118 1727204094.47250: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000088 13118 1727204094.47254: WORKER PROCESS EXITING ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13118 1727204094.47311: no more pending results, returning what we have 13118 1727204094.47315: results queue empty 13118 1727204094.47316: checking for any_errors_fatal 13118 1727204094.47322: done checking for any_errors_fatal 13118 1727204094.47323: checking for max_fail_percentage 13118 1727204094.47324: done checking for max_fail_percentage 13118 1727204094.47325: checking to see if all hosts have failed and the running result is not ok 13118 1727204094.47326: done checking to see if all hosts have failed 13118 1727204094.47327: getting the remaining hosts for this loop 13118 1727204094.47328: done getting the remaining hosts for this loop 13118 1727204094.47334: getting the next task for host managed-node2 13118 1727204094.47341: done getting next task for host managed-node2 13118 1727204094.47345: ^ task is: TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant 13118 1727204094.47348: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=17, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204094.47359: getting variables 13118 1727204094.47360: in VariableManager get_vars() 13118 1727204094.47397: Calling all_inventory to load vars for managed-node2 13118 1727204094.47399: Calling groups_inventory to load vars for managed-node2 13118 1727204094.47401: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204094.47411: Calling all_plugins_play to load vars for managed-node2 13118 1727204094.47413: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204094.47416: Calling groups_plugins_play to load vars for managed-node2 13118 1727204094.48851: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204094.50252: done with get_vars() 13118 1727204094.50284: done getting variables 13118 1727204094.50550: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable and start wpa_supplicant] ***** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:133 Tuesday 24 September 2024 14:54:54 -0400 (0:00:00.822) 0:00:31.806 ***** 13118 1727204094.50606: entering _queue_task() for managed-node2/service 13118 1727204094.51123: worker is 1 (out of 1 available) 13118 1727204094.51140: exiting _queue_task() for managed-node2/service 13118 1727204094.51151: done queuing things up, now waiting for results queue to drain 13118 1727204094.51153: waiting for pending results... 13118 1727204094.51478: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant 13118 1727204094.51640: in run() - task 0affcd87-79f5-56a3-0a64-000000000089 13118 1727204094.51692: variable 'ansible_search_path' from source: unknown 13118 1727204094.51703: variable 'ansible_search_path' from source: unknown 13118 1727204094.51761: calling self._execute() 13118 1727204094.51877: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204094.51889: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204094.51903: variable 'omit' from source: magic vars 13118 1727204094.52308: variable 'ansible_distribution_major_version' from source: facts 13118 1727204094.52327: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204094.52455: variable 'network_provider' from source: set_fact 13118 1727204094.52471: Evaluated conditional (network_provider == "nm"): True 13118 1727204094.52567: variable '__network_wpa_supplicant_required' from source: role '' defaults 13118 1727204094.52739: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13118 1727204094.52942: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204094.56778: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204094.56782: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204094.56785: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204094.56788: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204094.56790: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204094.56801: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204094.56869: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204094.56910: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204094.56951: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204094.57173: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204094.57176: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204094.57179: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204094.57182: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204094.57184: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204094.57186: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204094.57188: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204094.57190: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204094.57218: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204094.57574: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204094.57577: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204094.57580: variable 'network_connections' from source: task vars 13118 1727204094.57582: variable 'port2_profile' from source: play vars 13118 1727204094.57585: variable 'port2_profile' from source: play vars 13118 1727204094.57587: variable 'port1_profile' from source: play vars 13118 1727204094.57690: variable 'port1_profile' from source: play vars 13118 1727204094.57693: variable 'controller_profile' from source: play vars 13118 1727204094.57770: variable 'controller_profile' from source: play vars 13118 1727204094.57882: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13118 1727204094.58073: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13118 1727204094.58076: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13118 1727204094.58183: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13118 1727204094.58186: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13118 1727204094.58244: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13118 1727204094.58250: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13118 1727204094.58254: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204094.58284: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13118 1727204094.58343: variable '__network_wireless_connections_defined' from source: role '' defaults 13118 1727204094.58668: variable 'network_connections' from source: task vars 13118 1727204094.58683: variable 'port2_profile' from source: play vars 13118 1727204094.58761: variable 'port2_profile' from source: play vars 13118 1727204094.58775: variable 'port1_profile' from source: play vars 13118 1727204094.58861: variable 'port1_profile' from source: play vars 13118 1727204094.58885: variable 'controller_profile' from source: play vars 13118 1727204094.58969: variable 'controller_profile' from source: play vars 13118 1727204094.59003: Evaluated conditional (__network_wpa_supplicant_required): False 13118 1727204094.59011: when evaluation is False, skipping this task 13118 1727204094.59017: _execute() done 13118 1727204094.59023: dumping result to json 13118 1727204094.59044: done dumping result, returning 13118 1727204094.59061: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant [0affcd87-79f5-56a3-0a64-000000000089] 13118 1727204094.59081: sending task result for task 0affcd87-79f5-56a3-0a64-000000000089 skipping: [managed-node2] => { "changed": false, "false_condition": "__network_wpa_supplicant_required", "skip_reason": "Conditional result was False" } 13118 1727204094.59250: no more pending results, returning what we have 13118 1727204094.59255: results queue empty 13118 1727204094.59256: checking for any_errors_fatal 13118 1727204094.59280: done checking for any_errors_fatal 13118 1727204094.59281: checking for max_fail_percentage 13118 1727204094.59283: done checking for max_fail_percentage 13118 1727204094.59284: checking to see if all hosts have failed and the running result is not ok 13118 1727204094.59285: done checking to see if all hosts have failed 13118 1727204094.59286: getting the remaining hosts for this loop 13118 1727204094.59287: done getting the remaining hosts for this loop 13118 1727204094.59291: getting the next task for host managed-node2 13118 1727204094.59299: done getting next task for host managed-node2 13118 1727204094.59303: ^ task is: TASK: fedora.linux_system_roles.network : Enable network service 13118 1727204094.59308: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204094.59327: getting variables 13118 1727204094.59329: in VariableManager get_vars() 13118 1727204094.59378: Calling all_inventory to load vars for managed-node2 13118 1727204094.59381: Calling groups_inventory to load vars for managed-node2 13118 1727204094.59384: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204094.59394: Calling all_plugins_play to load vars for managed-node2 13118 1727204094.59398: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204094.59401: Calling groups_plugins_play to load vars for managed-node2 13118 1727204094.60524: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000089 13118 1727204094.60528: WORKER PROCESS EXITING 13118 1727204094.61627: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204094.62619: done with get_vars() 13118 1727204094.62639: done getting variables 13118 1727204094.62691: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable network service] ************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:142 Tuesday 24 September 2024 14:54:54 -0400 (0:00:00.121) 0:00:31.927 ***** 13118 1727204094.62716: entering _queue_task() for managed-node2/service 13118 1727204094.62955: worker is 1 (out of 1 available) 13118 1727204094.62972: exiting _queue_task() for managed-node2/service 13118 1727204094.62984: done queuing things up, now waiting for results queue to drain 13118 1727204094.62986: waiting for pending results... 13118 1727204094.63274: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Enable network service 13118 1727204094.63438: in run() - task 0affcd87-79f5-56a3-0a64-00000000008a 13118 1727204094.63470: variable 'ansible_search_path' from source: unknown 13118 1727204094.63481: variable 'ansible_search_path' from source: unknown 13118 1727204094.63522: calling self._execute() 13118 1727204094.63643: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204094.63657: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204094.63678: variable 'omit' from source: magic vars 13118 1727204094.64502: variable 'ansible_distribution_major_version' from source: facts 13118 1727204094.64520: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204094.64654: variable 'network_provider' from source: set_fact 13118 1727204094.64667: Evaluated conditional (network_provider == "initscripts"): False 13118 1727204094.64676: when evaluation is False, skipping this task 13118 1727204094.64692: _execute() done 13118 1727204094.64701: dumping result to json 13118 1727204094.64708: done dumping result, returning 13118 1727204094.64719: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Enable network service [0affcd87-79f5-56a3-0a64-00000000008a] 13118 1727204094.64734: sending task result for task 0affcd87-79f5-56a3-0a64-00000000008a skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13118 1727204094.64910: no more pending results, returning what we have 13118 1727204094.64914: results queue empty 13118 1727204094.64915: checking for any_errors_fatal 13118 1727204094.64923: done checking for any_errors_fatal 13118 1727204094.64923: checking for max_fail_percentage 13118 1727204094.64925: done checking for max_fail_percentage 13118 1727204094.64926: checking to see if all hosts have failed and the running result is not ok 13118 1727204094.64927: done checking to see if all hosts have failed 13118 1727204094.64927: getting the remaining hosts for this loop 13118 1727204094.64929: done getting the remaining hosts for this loop 13118 1727204094.64932: getting the next task for host managed-node2 13118 1727204094.64940: done getting next task for host managed-node2 13118 1727204094.64944: ^ task is: TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present 13118 1727204094.64948: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=19, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204094.64976: getting variables 13118 1727204094.64978: in VariableManager get_vars() 13118 1727204094.65019: Calling all_inventory to load vars for managed-node2 13118 1727204094.65022: Calling groups_inventory to load vars for managed-node2 13118 1727204094.65024: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204094.65037: Calling all_plugins_play to load vars for managed-node2 13118 1727204094.65040: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204094.65043: Calling groups_plugins_play to load vars for managed-node2 13118 1727204094.65560: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000008a 13118 1727204094.65564: WORKER PROCESS EXITING 13118 1727204094.66053: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204094.67356: done with get_vars() 13118 1727204094.67388: done getting variables 13118 1727204094.67458: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Ensure initscripts network file dependency is present] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:150 Tuesday 24 September 2024 14:54:54 -0400 (0:00:00.047) 0:00:31.974 ***** 13118 1727204094.67488: entering _queue_task() for managed-node2/copy 13118 1727204094.67839: worker is 1 (out of 1 available) 13118 1727204094.67856: exiting _queue_task() for managed-node2/copy 13118 1727204094.67872: done queuing things up, now waiting for results queue to drain 13118 1727204094.67874: waiting for pending results... 13118 1727204094.68100: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present 13118 1727204094.68202: in run() - task 0affcd87-79f5-56a3-0a64-00000000008b 13118 1727204094.68214: variable 'ansible_search_path' from source: unknown 13118 1727204094.68218: variable 'ansible_search_path' from source: unknown 13118 1727204094.68251: calling self._execute() 13118 1727204094.68329: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204094.68332: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204094.68346: variable 'omit' from source: magic vars 13118 1727204094.68773: variable 'ansible_distribution_major_version' from source: facts 13118 1727204094.68799: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204094.68903: variable 'network_provider' from source: set_fact 13118 1727204094.68916: Evaluated conditional (network_provider == "initscripts"): False 13118 1727204094.68924: when evaluation is False, skipping this task 13118 1727204094.68932: _execute() done 13118 1727204094.68940: dumping result to json 13118 1727204094.68948: done dumping result, returning 13118 1727204094.68969: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present [0affcd87-79f5-56a3-0a64-00000000008b] 13118 1727204094.68982: sending task result for task 0affcd87-79f5-56a3-0a64-00000000008b skipping: [managed-node2] => { "changed": false, "false_condition": "network_provider == \"initscripts\"", "skip_reason": "Conditional result was False" } 13118 1727204094.69241: no more pending results, returning what we have 13118 1727204094.69246: results queue empty 13118 1727204094.69250: checking for any_errors_fatal 13118 1727204094.69258: done checking for any_errors_fatal 13118 1727204094.69259: checking for max_fail_percentage 13118 1727204094.69261: done checking for max_fail_percentage 13118 1727204094.69262: checking to see if all hosts have failed and the running result is not ok 13118 1727204094.69263: done checking to see if all hosts have failed 13118 1727204094.69265: getting the remaining hosts for this loop 13118 1727204094.69267: done getting the remaining hosts for this loop 13118 1727204094.69271: getting the next task for host managed-node2 13118 1727204094.69279: done getting next task for host managed-node2 13118 1727204094.69283: ^ task is: TASK: fedora.linux_system_roles.network : Configure networking connection profiles 13118 1727204094.69287: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=20, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204094.69298: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000008b 13118 1727204094.69300: WORKER PROCESS EXITING 13118 1727204094.69316: getting variables 13118 1727204094.69321: in VariableManager get_vars() 13118 1727204094.69379: Calling all_inventory to load vars for managed-node2 13118 1727204094.69382: Calling groups_inventory to load vars for managed-node2 13118 1727204094.69388: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204094.69406: Calling all_plugins_play to load vars for managed-node2 13118 1727204094.69410: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204094.69414: Calling groups_plugins_play to load vars for managed-node2 13118 1727204094.70904: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204094.72253: done with get_vars() 13118 1727204094.72280: done getting variables TASK [fedora.linux_system_roles.network : Configure networking connection profiles] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:159 Tuesday 24 September 2024 14:54:54 -0400 (0:00:00.048) 0:00:32.023 ***** 13118 1727204094.72351: entering _queue_task() for managed-node2/fedora.linux_system_roles.network_connections 13118 1727204094.72660: worker is 1 (out of 1 available) 13118 1727204094.72675: exiting _queue_task() for managed-node2/fedora.linux_system_roles.network_connections 13118 1727204094.72686: done queuing things up, now waiting for results queue to drain 13118 1727204094.72687: waiting for pending results... 13118 1727204094.72884: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Configure networking connection profiles 13118 1727204094.73074: in run() - task 0affcd87-79f5-56a3-0a64-00000000008c 13118 1727204094.73126: variable 'ansible_search_path' from source: unknown 13118 1727204094.73138: variable 'ansible_search_path' from source: unknown 13118 1727204094.73183: calling self._execute() 13118 1727204094.73303: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204094.73319: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204094.73352: variable 'omit' from source: magic vars 13118 1727204094.73816: variable 'ansible_distribution_major_version' from source: facts 13118 1727204094.73869: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204094.73909: variable 'omit' from source: magic vars 13118 1727204094.74019: variable 'omit' from source: magic vars 13118 1727204094.74176: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13118 1727204094.75815: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13118 1727204094.75906: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13118 1727204094.75995: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13118 1727204094.76051: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13118 1727204094.76094: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13118 1727204094.76203: variable 'network_provider' from source: set_fact 13118 1727204094.76373: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13118 1727204094.76456: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13118 1727204094.76505: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13118 1727204094.76562: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13118 1727204094.76592: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13118 1727204094.76691: variable 'omit' from source: magic vars 13118 1727204094.76852: variable 'omit' from source: magic vars 13118 1727204094.76981: variable 'network_connections' from source: task vars 13118 1727204094.77135: variable 'port2_profile' from source: play vars 13118 1727204094.77225: variable 'port2_profile' from source: play vars 13118 1727204094.77243: variable 'port1_profile' from source: play vars 13118 1727204094.77328: variable 'port1_profile' from source: play vars 13118 1727204094.77352: variable 'controller_profile' from source: play vars 13118 1727204094.77438: variable 'controller_profile' from source: play vars 13118 1727204094.77645: variable 'omit' from source: magic vars 13118 1727204094.77669: variable '__lsr_ansible_managed' from source: task vars 13118 1727204094.77743: variable '__lsr_ansible_managed' from source: task vars 13118 1727204094.77969: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup 13118 1727204094.78258: Loaded config def from plugin (lookup/template) 13118 1727204094.78271: Loading LookupModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup/template.py 13118 1727204094.78312: File lookup term: get_ansible_managed.j2 13118 1727204094.78322: variable 'ansible_search_path' from source: unknown 13118 1727204094.78344: evaluation_path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks 13118 1727204094.78368: search_path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/templates/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/templates/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/templates/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/templates/get_ansible_managed.j2 /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/get_ansible_managed.j2 13118 1727204094.78400: variable 'ansible_search_path' from source: unknown 13118 1727204094.84260: variable 'ansible_managed' from source: unknown 13118 1727204094.84373: variable 'omit' from source: magic vars 13118 1727204094.84395: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204094.84422: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204094.84449: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204094.84463: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204094.84474: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204094.84514: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204094.84539: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204094.84544: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204094.84658: Set connection var ansible_timeout to 10 13118 1727204094.84681: Set connection var ansible_pipelining to False 13118 1727204094.84684: Set connection var ansible_connection to ssh 13118 1727204094.84692: Set connection var ansible_shell_executable to /bin/sh 13118 1727204094.84694: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204094.84697: Set connection var ansible_shell_type to sh 13118 1727204094.84730: variable 'ansible_shell_executable' from source: unknown 13118 1727204094.84738: variable 'ansible_connection' from source: unknown 13118 1727204094.84740: variable 'ansible_module_compression' from source: unknown 13118 1727204094.84744: variable 'ansible_shell_type' from source: unknown 13118 1727204094.84746: variable 'ansible_shell_executable' from source: unknown 13118 1727204094.84749: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204094.84751: variable 'ansible_pipelining' from source: unknown 13118 1727204094.84753: variable 'ansible_timeout' from source: unknown 13118 1727204094.84774: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204094.84870: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204094.84881: variable 'omit' from source: magic vars 13118 1727204094.84899: starting attempt loop 13118 1727204094.84912: running the handler 13118 1727204094.84920: _low_level_execute_command(): starting 13118 1727204094.84927: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204094.85884: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204094.85897: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.85924: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.85990: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.86029: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204094.86040: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204094.86066: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.86087: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204094.86130: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204094.86145: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204094.86167: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.86178: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.86190: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.86198: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204094.86212: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204094.86234: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.86368: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204094.86372: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204094.86375: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204094.86479: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204094.88126: stdout chunk (state=3): >>>/root <<< 13118 1727204094.88255: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204094.88315: stderr chunk (state=3): >>><<< 13118 1727204094.88319: stdout chunk (state=3): >>><<< 13118 1727204094.88347: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204094.88385: _low_level_execute_command(): starting 13118 1727204094.88388: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204094.8836002-15717-244963618923648 `" && echo ansible-tmp-1727204094.8836002-15717-244963618923648="` echo /root/.ansible/tmp/ansible-tmp-1727204094.8836002-15717-244963618923648 `" ) && sleep 0' 13118 1727204094.89186: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204094.89223: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.89251: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.89255: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.89448: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204094.89452: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204094.89492: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204094.91343: stdout chunk (state=3): >>>ansible-tmp-1727204094.8836002-15717-244963618923648=/root/.ansible/tmp/ansible-tmp-1727204094.8836002-15717-244963618923648 <<< 13118 1727204094.91488: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204094.91510: stderr chunk (state=3): >>><<< 13118 1727204094.91513: stdout chunk (state=3): >>><<< 13118 1727204094.91533: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204094.8836002-15717-244963618923648=/root/.ansible/tmp/ansible-tmp-1727204094.8836002-15717-244963618923648 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204094.91576: variable 'ansible_module_compression' from source: unknown 13118 1727204094.91617: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible_collections.fedora.linux_system_roles.plugins.modules.network_connections-ZIP_DEFLATED 13118 1727204094.91647: variable 'ansible_facts' from source: unknown 13118 1727204094.91716: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204094.8836002-15717-244963618923648/AnsiballZ_network_connections.py 13118 1727204094.91833: Sending initial data 13118 1727204094.91841: Sent initial data (168 bytes) 13118 1727204094.92719: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204094.92728: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.92736: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.92747: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.92776: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204094.92783: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204094.92793: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.92801: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204094.92809: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204094.92814: stderr chunk (state=3): >>>debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.92823: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.92834: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.92843: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204094.92848: stderr chunk (state=3): >>>debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.92899: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204094.92938: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204094.92948: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204094.93080: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204094.94682: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204094.94717: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204094.94755: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmp1hp1s83s /root/.ansible/tmp/ansible-tmp-1727204094.8836002-15717-244963618923648/AnsiballZ_network_connections.py <<< 13118 1727204094.94807: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204094.96484: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204094.96690: stderr chunk (state=3): >>><<< 13118 1727204094.96693: stdout chunk (state=3): >>><<< 13118 1727204094.96695: done transferring module to remote 13118 1727204094.96697: _low_level_execute_command(): starting 13118 1727204094.96699: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204094.8836002-15717-244963618923648/ /root/.ansible/tmp/ansible-tmp-1727204094.8836002-15717-244963618923648/AnsiballZ_network_connections.py && sleep 0' 13118 1727204094.97637: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204094.97652: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.97669: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.97699: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.97759: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204094.97773: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204094.97786: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.97808: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204094.97823: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204094.97833: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204094.97844: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204094.97857: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204094.97907: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204094.97951: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204094.98017: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204094.98057: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204094.98238: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204094.98293: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204094.98342: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204094.98447: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204095.00201: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204095.00325: stderr chunk (state=3): >>><<< 13118 1727204095.00352: stdout chunk (state=3): >>><<< 13118 1727204095.00479: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204095.00483: _low_level_execute_command(): starting 13118 1727204095.00486: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204094.8836002-15717-244963618923648/AnsiballZ_network_connections.py && sleep 0' 13118 1727204095.01212: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204095.01228: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204095.01251: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204095.01276: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204095.01319: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204095.01331: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204095.01352: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204095.01378: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204095.01390: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204095.01401: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204095.01413: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204095.01425: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204095.01440: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204095.01451: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204095.01471: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204095.01490: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204095.01575: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204095.01658: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204095.01724: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204095.01816: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204095.50573: stdout chunk (state=3): >>>Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_94fxt9wo/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_94fxt9wo/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0.1/161d3351-eab4-440b-a3b9-ee4281b1f747: error=unknown <<< 13118 1727204095.52488: stdout chunk (state=3): >>>Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_94fxt9wo/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_94fxt9wo/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail <<< 13118 1727204095.52509: stdout chunk (state=3): >>>ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0.0/382c36b9-9f8f-4077-be2c-b77f342639e8: error=unknown <<< 13118 1727204095.54344: stdout chunk (state=3): >>>Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_94fxt9wo/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back <<< 13118 1727204095.54360: stdout chunk (state=3): >>> File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_94fxt9wo/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail <<< 13118 1727204095.54383: stdout chunk (state=3): >>>ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0/debabe6d-f673-439e-9f14-25954ecc55a3: error=unknown <<< 13118 1727204095.54643: stdout chunk (state=3): >>> {"changed": true, "warnings": [], "stderr": "\n", "_invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0.1", "persistent_state": "absent", "state": "down"}, {"name": "bond0.0", "persistent_state": "absent", "state": "down"}, {"name": "bond0", "persistent_state": "absent", "state": "down"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}, "invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0.1", "persistent_state": "absent", "state": "down"}, {"name": "bond0.0", "persistent_state": "absent", "state": "down"}, {"name": "bond0", "persistent_state": "absent", "state": "down"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}} <<< 13118 1727204095.56414: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204095.56418: stdout chunk (state=3): >>><<< 13118 1727204095.56420: stderr chunk (state=3): >>><<< 13118 1727204095.56471: _low_level_execute_command() done: rc=0, stdout=Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_94fxt9wo/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_94fxt9wo/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0.1/161d3351-eab4-440b-a3b9-ee4281b1f747: error=unknown Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_94fxt9wo/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_94fxt9wo/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0.0/382c36b9-9f8f-4077-be2c-b77f342639e8: error=unknown Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_94fxt9wo/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_94fxt9wo/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0/debabe6d-f673-439e-9f14-25954ecc55a3: error=unknown {"changed": true, "warnings": [], "stderr": "\n", "_invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0.1", "persistent_state": "absent", "state": "down"}, {"name": "bond0.0", "persistent_state": "absent", "state": "down"}, {"name": "bond0", "persistent_state": "absent", "state": "down"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}, "invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0.1", "persistent_state": "absent", "state": "down"}, {"name": "bond0.0", "persistent_state": "absent", "state": "down"}, {"name": "bond0", "persistent_state": "absent", "state": "down"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204095.56670: done with _execute_module (fedora.linux_system_roles.network_connections, {'provider': 'nm', 'connections': [{'name': 'bond0.1', 'persistent_state': 'absent', 'state': 'down'}, {'name': 'bond0.0', 'persistent_state': 'absent', 'state': 'down'}, {'name': 'bond0', 'persistent_state': 'absent', 'state': 'down'}], '__header': '#\n# Ansible managed\n#\n# system_role:network\n', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'fedora.linux_system_roles.network_connections', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204094.8836002-15717-244963618923648/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204095.56674: _low_level_execute_command(): starting 13118 1727204095.56678: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204094.8836002-15717-244963618923648/ > /dev/null 2>&1 && sleep 0' 13118 1727204095.57383: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204095.57400: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204095.57417: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204095.57440: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204095.57492: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204095.57504: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204095.57516: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204095.57535: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204095.57548: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204095.57573: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204095.57587: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204095.57602: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204095.57620: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204095.57635: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204095.57647: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204095.57710: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204095.57915: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204095.57940: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204095.57955: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204095.58508: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204095.60405: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204095.60619: stderr chunk (state=3): >>><<< 13118 1727204095.60623: stdout chunk (state=3): >>><<< 13118 1727204095.60878: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204095.60881: handler run complete 13118 1727204095.60883: attempt loop complete, returning result 13118 1727204095.60885: _execute() done 13118 1727204095.60887: dumping result to json 13118 1727204095.60889: done dumping result, returning 13118 1727204095.60891: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Configure networking connection profiles [0affcd87-79f5-56a3-0a64-00000000008c] 13118 1727204095.60893: sending task result for task 0affcd87-79f5-56a3-0a64-00000000008c 13118 1727204095.61202: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000008c 13118 1727204095.61206: WORKER PROCESS EXITING changed: [managed-node2] => { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "name": "bond0.1", "persistent_state": "absent", "state": "down" }, { "name": "bond0.0", "persistent_state": "absent", "state": "down" }, { "name": "bond0", "persistent_state": "absent", "state": "down" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true } STDERR: 13118 1727204095.61333: no more pending results, returning what we have 13118 1727204095.61337: results queue empty 13118 1727204095.61339: checking for any_errors_fatal 13118 1727204095.61345: done checking for any_errors_fatal 13118 1727204095.61345: checking for max_fail_percentage 13118 1727204095.61348: done checking for max_fail_percentage 13118 1727204095.61349: checking to see if all hosts have failed and the running result is not ok 13118 1727204095.61350: done checking to see if all hosts have failed 13118 1727204095.61350: getting the remaining hosts for this loop 13118 1727204095.61351: done getting the remaining hosts for this loop 13118 1727204095.61355: getting the next task for host managed-node2 13118 1727204095.61372: done getting next task for host managed-node2 13118 1727204095.61376: ^ task is: TASK: fedora.linux_system_roles.network : Configure networking state 13118 1727204095.61380: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=21, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204095.61398: getting variables 13118 1727204095.61400: in VariableManager get_vars() 13118 1727204095.61441: Calling all_inventory to load vars for managed-node2 13118 1727204095.61444: Calling groups_inventory to load vars for managed-node2 13118 1727204095.61447: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204095.61457: Calling all_plugins_play to load vars for managed-node2 13118 1727204095.61460: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204095.61465: Calling groups_plugins_play to load vars for managed-node2 13118 1727204095.64260: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204095.66290: done with get_vars() 13118 1727204095.66330: done getting variables TASK [fedora.linux_system_roles.network : Configure networking state] ********** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:171 Tuesday 24 September 2024 14:54:55 -0400 (0:00:00.940) 0:00:32.964 ***** 13118 1727204095.66438: entering _queue_task() for managed-node2/fedora.linux_system_roles.network_state 13118 1727204095.66813: worker is 1 (out of 1 available) 13118 1727204095.66829: exiting _queue_task() for managed-node2/fedora.linux_system_roles.network_state 13118 1727204095.66841: done queuing things up, now waiting for results queue to drain 13118 1727204095.66848: waiting for pending results... 13118 1727204095.67170: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Configure networking state 13118 1727204095.67339: in run() - task 0affcd87-79f5-56a3-0a64-00000000008d 13118 1727204095.67369: variable 'ansible_search_path' from source: unknown 13118 1727204095.67381: variable 'ansible_search_path' from source: unknown 13118 1727204095.67527: calling self._execute() 13118 1727204095.67767: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204095.67780: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204095.67792: variable 'omit' from source: magic vars 13118 1727204095.68215: variable 'ansible_distribution_major_version' from source: facts 13118 1727204095.68234: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204095.68386: variable 'network_state' from source: role '' defaults 13118 1727204095.68410: Evaluated conditional (network_state != {}): False 13118 1727204095.68417: when evaluation is False, skipping this task 13118 1727204095.68422: _execute() done 13118 1727204095.68428: dumping result to json 13118 1727204095.68433: done dumping result, returning 13118 1727204095.68442: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Configure networking state [0affcd87-79f5-56a3-0a64-00000000008d] 13118 1727204095.68456: sending task result for task 0affcd87-79f5-56a3-0a64-00000000008d skipping: [managed-node2] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13118 1727204095.68634: no more pending results, returning what we have 13118 1727204095.68639: results queue empty 13118 1727204095.68640: checking for any_errors_fatal 13118 1727204095.68653: done checking for any_errors_fatal 13118 1727204095.68654: checking for max_fail_percentage 13118 1727204095.68656: done checking for max_fail_percentage 13118 1727204095.68657: checking to see if all hosts have failed and the running result is not ok 13118 1727204095.68658: done checking to see if all hosts have failed 13118 1727204095.68659: getting the remaining hosts for this loop 13118 1727204095.68660: done getting the remaining hosts for this loop 13118 1727204095.68666: getting the next task for host managed-node2 13118 1727204095.68677: done getting next task for host managed-node2 13118 1727204095.68682: ^ task is: TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections 13118 1727204095.68688: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=22, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204095.68711: getting variables 13118 1727204095.68713: in VariableManager get_vars() 13118 1727204095.68755: Calling all_inventory to load vars for managed-node2 13118 1727204095.68758: Calling groups_inventory to load vars for managed-node2 13118 1727204095.68760: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204095.68774: Calling all_plugins_play to load vars for managed-node2 13118 1727204095.68777: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204095.68780: Calling groups_plugins_play to load vars for managed-node2 13118 1727204095.69875: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000008d 13118 1727204095.69879: WORKER PROCESS EXITING 13118 1727204095.71178: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204095.73444: done with get_vars() 13118 1727204095.73469: done getting variables 13118 1727204095.73513: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show stderr messages for the network_connections] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:177 Tuesday 24 September 2024 14:54:55 -0400 (0:00:00.071) 0:00:33.035 ***** 13118 1727204095.73546: entering _queue_task() for managed-node2/debug 13118 1727204095.73789: worker is 1 (out of 1 available) 13118 1727204095.73803: exiting _queue_task() for managed-node2/debug 13118 1727204095.73814: done queuing things up, now waiting for results queue to drain 13118 1727204095.73816: waiting for pending results... 13118 1727204095.74014: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections 13118 1727204095.74125: in run() - task 0affcd87-79f5-56a3-0a64-00000000008e 13118 1727204095.74139: variable 'ansible_search_path' from source: unknown 13118 1727204095.74144: variable 'ansible_search_path' from source: unknown 13118 1727204095.74175: calling self._execute() 13118 1727204095.74252: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204095.74257: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204095.74266: variable 'omit' from source: magic vars 13118 1727204095.74542: variable 'ansible_distribution_major_version' from source: facts 13118 1727204095.74552: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204095.74557: variable 'omit' from source: magic vars 13118 1727204095.74605: variable 'omit' from source: magic vars 13118 1727204095.74636: variable 'omit' from source: magic vars 13118 1727204095.74671: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204095.74698: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204095.74716: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204095.74735: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204095.74744: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204095.74769: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204095.74772: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204095.74774: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204095.74847: Set connection var ansible_timeout to 10 13118 1727204095.74856: Set connection var ansible_pipelining to False 13118 1727204095.74859: Set connection var ansible_connection to ssh 13118 1727204095.74863: Set connection var ansible_shell_executable to /bin/sh 13118 1727204095.74875: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204095.74877: Set connection var ansible_shell_type to sh 13118 1727204095.74891: variable 'ansible_shell_executable' from source: unknown 13118 1727204095.74894: variable 'ansible_connection' from source: unknown 13118 1727204095.74896: variable 'ansible_module_compression' from source: unknown 13118 1727204095.74898: variable 'ansible_shell_type' from source: unknown 13118 1727204095.74901: variable 'ansible_shell_executable' from source: unknown 13118 1727204095.74903: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204095.74908: variable 'ansible_pipelining' from source: unknown 13118 1727204095.74911: variable 'ansible_timeout' from source: unknown 13118 1727204095.74913: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204095.75015: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204095.75024: variable 'omit' from source: magic vars 13118 1727204095.75030: starting attempt loop 13118 1727204095.75037: running the handler 13118 1727204095.75129: variable '__network_connections_result' from source: set_fact 13118 1727204095.75179: handler run complete 13118 1727204095.75191: attempt loop complete, returning result 13118 1727204095.75194: _execute() done 13118 1727204095.75196: dumping result to json 13118 1727204095.75198: done dumping result, returning 13118 1727204095.75206: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections [0affcd87-79f5-56a3-0a64-00000000008e] 13118 1727204095.75221: sending task result for task 0affcd87-79f5-56a3-0a64-00000000008e 13118 1727204095.75353: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000008e ok: [managed-node2] => { "__network_connections_result.stderr_lines": [ "" ] } 13118 1727204095.75438: no more pending results, returning what we have 13118 1727204095.75441: results queue empty 13118 1727204095.75442: checking for any_errors_fatal 13118 1727204095.75448: done checking for any_errors_fatal 13118 1727204095.75564: checking for max_fail_percentage 13118 1727204095.75567: done checking for max_fail_percentage 13118 1727204095.75569: checking to see if all hosts have failed and the running result is not ok 13118 1727204095.75569: done checking to see if all hosts have failed 13118 1727204095.75570: getting the remaining hosts for this loop 13118 1727204095.75575: done getting the remaining hosts for this loop 13118 1727204095.75579: getting the next task for host managed-node2 13118 1727204095.75591: done getting next task for host managed-node2 13118 1727204095.75595: ^ task is: TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections 13118 1727204095.75600: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=23, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204095.75610: WORKER PROCESS EXITING 13118 1727204095.75619: getting variables 13118 1727204095.75621: in VariableManager get_vars() 13118 1727204095.75770: Calling all_inventory to load vars for managed-node2 13118 1727204095.75773: Calling groups_inventory to load vars for managed-node2 13118 1727204095.75785: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204095.75800: Calling all_plugins_play to load vars for managed-node2 13118 1727204095.75808: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204095.75812: Calling groups_plugins_play to load vars for managed-node2 13118 1727204095.77751: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204095.79313: done with get_vars() 13118 1727204095.79348: done getting variables 13118 1727204095.79414: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show debug messages for the network_connections] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:181 Tuesday 24 September 2024 14:54:55 -0400 (0:00:00.059) 0:00:33.094 ***** 13118 1727204095.79457: entering _queue_task() for managed-node2/debug 13118 1727204095.79945: worker is 1 (out of 1 available) 13118 1727204095.79959: exiting _queue_task() for managed-node2/debug 13118 1727204095.79974: done queuing things up, now waiting for results queue to drain 13118 1727204095.79976: waiting for pending results... 13118 1727204095.80199: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections 13118 1727204095.80332: in run() - task 0affcd87-79f5-56a3-0a64-00000000008f 13118 1727204095.80345: variable 'ansible_search_path' from source: unknown 13118 1727204095.80349: variable 'ansible_search_path' from source: unknown 13118 1727204095.80407: calling self._execute() 13118 1727204095.80492: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204095.80515: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204095.80518: variable 'omit' from source: magic vars 13118 1727204095.80908: variable 'ansible_distribution_major_version' from source: facts 13118 1727204095.80921: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204095.80925: variable 'omit' from source: magic vars 13118 1727204095.81412: variable 'omit' from source: magic vars 13118 1727204095.81618: variable 'omit' from source: magic vars 13118 1727204095.81873: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204095.82242: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204095.82268: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204095.82284: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204095.82298: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204095.82327: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204095.82332: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204095.82335: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204095.82558: Set connection var ansible_timeout to 10 13118 1727204095.82572: Set connection var ansible_pipelining to False 13118 1727204095.82575: Set connection var ansible_connection to ssh 13118 1727204095.82580: Set connection var ansible_shell_executable to /bin/sh 13118 1727204095.82586: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204095.82589: Set connection var ansible_shell_type to sh 13118 1727204095.82614: variable 'ansible_shell_executable' from source: unknown 13118 1727204095.82618: variable 'ansible_connection' from source: unknown 13118 1727204095.82621: variable 'ansible_module_compression' from source: unknown 13118 1727204095.82624: variable 'ansible_shell_type' from source: unknown 13118 1727204095.82626: variable 'ansible_shell_executable' from source: unknown 13118 1727204095.82628: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204095.82633: variable 'ansible_pipelining' from source: unknown 13118 1727204095.82636: variable 'ansible_timeout' from source: unknown 13118 1727204095.82638: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204095.82804: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204095.82815: variable 'omit' from source: magic vars 13118 1727204095.82821: starting attempt loop 13118 1727204095.82825: running the handler 13118 1727204095.82890: variable '__network_connections_result' from source: set_fact 13118 1727204095.82967: variable '__network_connections_result' from source: set_fact 13118 1727204095.83128: handler run complete 13118 1727204095.83176: attempt loop complete, returning result 13118 1727204095.83179: _execute() done 13118 1727204095.83181: dumping result to json 13118 1727204095.83183: done dumping result, returning 13118 1727204095.83186: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections [0affcd87-79f5-56a3-0a64-00000000008f] 13118 1727204095.83188: sending task result for task 0affcd87-79f5-56a3-0a64-00000000008f 13118 1727204095.83372: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000008f 13118 1727204095.83375: WORKER PROCESS EXITING ok: [managed-node2] => { "__network_connections_result": { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "name": "bond0.1", "persistent_state": "absent", "state": "down" }, { "name": "bond0.0", "persistent_state": "absent", "state": "down" }, { "name": "bond0", "persistent_state": "absent", "state": "down" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true, "failed": false, "stderr": "\n", "stderr_lines": [ "" ] } } 13118 1727204095.83636: no more pending results, returning what we have 13118 1727204095.83640: results queue empty 13118 1727204095.83641: checking for any_errors_fatal 13118 1727204095.83646: done checking for any_errors_fatal 13118 1727204095.83647: checking for max_fail_percentage 13118 1727204095.83648: done checking for max_fail_percentage 13118 1727204095.83649: checking to see if all hosts have failed and the running result is not ok 13118 1727204095.83650: done checking to see if all hosts have failed 13118 1727204095.83651: getting the remaining hosts for this loop 13118 1727204095.83652: done getting the remaining hosts for this loop 13118 1727204095.83656: getting the next task for host managed-node2 13118 1727204095.83666: done getting next task for host managed-node2 13118 1727204095.83671: ^ task is: TASK: fedora.linux_system_roles.network : Show debug messages for the network_state 13118 1727204095.83679: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=24, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204095.83690: getting variables 13118 1727204095.83692: in VariableManager get_vars() 13118 1727204095.83728: Calling all_inventory to load vars for managed-node2 13118 1727204095.83731: Calling groups_inventory to load vars for managed-node2 13118 1727204095.83734: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204095.83747: Calling all_plugins_play to load vars for managed-node2 13118 1727204095.83756: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204095.83759: Calling groups_plugins_play to load vars for managed-node2 13118 1727204095.86139: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204095.87144: done with get_vars() 13118 1727204095.87172: done getting variables 13118 1727204095.87248: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show debug messages for the network_state] *** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:186 Tuesday 24 September 2024 14:54:55 -0400 (0:00:00.078) 0:00:33.172 ***** 13118 1727204095.87288: entering _queue_task() for managed-node2/debug 13118 1727204095.87652: worker is 1 (out of 1 available) 13118 1727204095.87662: exiting _queue_task() for managed-node2/debug 13118 1727204095.87676: done queuing things up, now waiting for results queue to drain 13118 1727204095.87678: waiting for pending results... 13118 1727204095.87986: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Show debug messages for the network_state 13118 1727204095.88162: in run() - task 0affcd87-79f5-56a3-0a64-000000000090 13118 1727204095.88187: variable 'ansible_search_path' from source: unknown 13118 1727204095.88199: variable 'ansible_search_path' from source: unknown 13118 1727204095.88243: calling self._execute() 13118 1727204095.88384: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204095.88396: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204095.88416: variable 'omit' from source: magic vars 13118 1727204095.88734: variable 'ansible_distribution_major_version' from source: facts 13118 1727204095.88741: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204095.88827: variable 'network_state' from source: role '' defaults 13118 1727204095.88837: Evaluated conditional (network_state != {}): False 13118 1727204095.88840: when evaluation is False, skipping this task 13118 1727204095.88842: _execute() done 13118 1727204095.88846: dumping result to json 13118 1727204095.88849: done dumping result, returning 13118 1727204095.88854: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Show debug messages for the network_state [0affcd87-79f5-56a3-0a64-000000000090] 13118 1727204095.88861: sending task result for task 0affcd87-79f5-56a3-0a64-000000000090 13118 1727204095.88952: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000090 13118 1727204095.88955: WORKER PROCESS EXITING skipping: [managed-node2] => { "false_condition": "network_state != {}" } 13118 1727204095.89039: no more pending results, returning what we have 13118 1727204095.89043: results queue empty 13118 1727204095.89044: checking for any_errors_fatal 13118 1727204095.89051: done checking for any_errors_fatal 13118 1727204095.89052: checking for max_fail_percentage 13118 1727204095.89054: done checking for max_fail_percentage 13118 1727204095.89055: checking to see if all hosts have failed and the running result is not ok 13118 1727204095.89055: done checking to see if all hosts have failed 13118 1727204095.89056: getting the remaining hosts for this loop 13118 1727204095.89057: done getting the remaining hosts for this loop 13118 1727204095.89060: getting the next task for host managed-node2 13118 1727204095.89068: done getting next task for host managed-node2 13118 1727204095.89072: ^ task is: TASK: fedora.linux_system_roles.network : Re-test connectivity 13118 1727204095.89076: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=25, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204095.89091: getting variables 13118 1727204095.89093: in VariableManager get_vars() 13118 1727204095.89124: Calling all_inventory to load vars for managed-node2 13118 1727204095.89126: Calling groups_inventory to load vars for managed-node2 13118 1727204095.89128: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204095.89140: Calling all_plugins_play to load vars for managed-node2 13118 1727204095.89143: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204095.89145: Calling groups_plugins_play to load vars for managed-node2 13118 1727204095.90004: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204095.93222: done with get_vars() 13118 1727204095.93253: done getting variables TASK [fedora.linux_system_roles.network : Re-test connectivity] **************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:192 Tuesday 24 September 2024 14:54:55 -0400 (0:00:00.061) 0:00:33.233 ***** 13118 1727204095.93399: entering _queue_task() for managed-node2/ping 13118 1727204095.93715: worker is 1 (out of 1 available) 13118 1727204095.93729: exiting _queue_task() for managed-node2/ping 13118 1727204095.93745: done queuing things up, now waiting for results queue to drain 13118 1727204095.93746: waiting for pending results... 13118 1727204095.94035: running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Re-test connectivity 13118 1727204095.94225: in run() - task 0affcd87-79f5-56a3-0a64-000000000091 13118 1727204095.94239: variable 'ansible_search_path' from source: unknown 13118 1727204095.94243: variable 'ansible_search_path' from source: unknown 13118 1727204095.94287: calling self._execute() 13118 1727204095.94369: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204095.94388: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204095.94418: variable 'omit' from source: magic vars 13118 1727204095.94903: variable 'ansible_distribution_major_version' from source: facts 13118 1727204095.94926: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204095.94938: variable 'omit' from source: magic vars 13118 1727204095.95021: variable 'omit' from source: magic vars 13118 1727204095.95066: variable 'omit' from source: magic vars 13118 1727204095.95121: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204095.95173: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204095.95197: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204095.95224: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204095.95255: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204095.95291: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204095.95300: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204095.95306: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204095.95429: Set connection var ansible_timeout to 10 13118 1727204095.95447: Set connection var ansible_pipelining to False 13118 1727204095.95453: Set connection var ansible_connection to ssh 13118 1727204095.95469: Set connection var ansible_shell_executable to /bin/sh 13118 1727204095.95481: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204095.95487: Set connection var ansible_shell_type to sh 13118 1727204095.95523: variable 'ansible_shell_executable' from source: unknown 13118 1727204095.95535: variable 'ansible_connection' from source: unknown 13118 1727204095.95542: variable 'ansible_module_compression' from source: unknown 13118 1727204095.95548: variable 'ansible_shell_type' from source: unknown 13118 1727204095.95553: variable 'ansible_shell_executable' from source: unknown 13118 1727204095.95558: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204095.95567: variable 'ansible_pipelining' from source: unknown 13118 1727204095.95577: variable 'ansible_timeout' from source: unknown 13118 1727204095.95587: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204095.95839: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) 13118 1727204095.95851: variable 'omit' from source: magic vars 13118 1727204095.95854: starting attempt loop 13118 1727204095.95869: running the handler 13118 1727204095.95909: _low_level_execute_command(): starting 13118 1727204095.95921: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204095.97215: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204095.97234: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204095.97242: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204095.97255: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204095.97297: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204095.97316: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204095.97319: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204095.97335: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204095.97360: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204095.97363: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204095.97414: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204095.97502: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204095.97540: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204095.97543: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204095.97554: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204095.97613: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204095.99269: stdout chunk (state=3): >>>/root <<< 13118 1727204095.99440: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204095.99632: stderr chunk (state=3): >>><<< 13118 1727204095.99641: stdout chunk (state=3): >>><<< 13118 1727204095.99673: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204095.99688: _low_level_execute_command(): starting 13118 1727204095.99694: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204095.9967358-15783-208253568161114 `" && echo ansible-tmp-1727204095.9967358-15783-208253568161114="` echo /root/.ansible/tmp/ansible-tmp-1727204095.9967358-15783-208253568161114 `" ) && sleep 0' 13118 1727204096.00770: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.00853: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.00857: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.00937: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.00969: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.01027: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.02926: stdout chunk (state=3): >>>ansible-tmp-1727204095.9967358-15783-208253568161114=/root/.ansible/tmp/ansible-tmp-1727204095.9967358-15783-208253568161114 <<< 13118 1727204096.03047: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204096.03109: stderr chunk (state=3): >>><<< 13118 1727204096.03112: stdout chunk (state=3): >>><<< 13118 1727204096.03127: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204095.9967358-15783-208253568161114=/root/.ansible/tmp/ansible-tmp-1727204095.9967358-15783-208253568161114 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204096.03177: variable 'ansible_module_compression' from source: unknown 13118 1727204096.03215: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.ping-ZIP_DEFLATED 13118 1727204096.03254: variable 'ansible_facts' from source: unknown 13118 1727204096.03320: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204095.9967358-15783-208253568161114/AnsiballZ_ping.py 13118 1727204096.03449: Sending initial data 13118 1727204096.03452: Sent initial data (153 bytes) 13118 1727204096.04309: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.04313: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.04323: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.04326: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204096.04339: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.04358: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204096.04362: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.04383: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.04395: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.04399: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.04407: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204096.04417: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.04482: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.04512: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.04552: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.06279: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204096.06319: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204096.06349: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpcgz43jy3 /root/.ansible/tmp/ansible-tmp-1727204095.9967358-15783-208253568161114/AnsiballZ_ping.py <<< 13118 1727204096.06388: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204096.07155: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204096.07269: stderr chunk (state=3): >>><<< 13118 1727204096.07272: stdout chunk (state=3): >>><<< 13118 1727204096.07289: done transferring module to remote 13118 1727204096.07298: _low_level_execute_command(): starting 13118 1727204096.07304: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204095.9967358-15783-208253568161114/ /root/.ansible/tmp/ansible-tmp-1727204095.9967358-15783-208253568161114/AnsiballZ_ping.py && sleep 0' 13118 1727204096.07845: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.07869: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.07905: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.07945: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.07948: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.08053: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.08059: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.08104: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.09841: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204096.09901: stderr chunk (state=3): >>><<< 13118 1727204096.09904: stdout chunk (state=3): >>><<< 13118 1727204096.09918: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204096.09921: _low_level_execute_command(): starting 13118 1727204096.09926: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204095.9967358-15783-208253568161114/AnsiballZ_ping.py && sleep 0' 13118 1727204096.10475: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.10480: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.10512: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.10516: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.10518: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.10576: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.10582: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204096.10584: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.10633: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.23647: stdout chunk (state=3): >>> {"ping": "pong", "invocation": {"module_args": {"data": "pong"}}} <<< 13118 1727204096.24734: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204096.24807: stderr chunk (state=3): >>><<< 13118 1727204096.24832: stdout chunk (state=3): >>><<< 13118 1727204096.24960: _low_level_execute_command() done: rc=0, stdout= {"ping": "pong", "invocation": {"module_args": {"data": "pong"}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204096.25493: done with _execute_module (ping, {'_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ping', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204095.9967358-15783-208253568161114/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204096.25668: _low_level_execute_command(): starting 13118 1727204096.25812: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204095.9967358-15783-208253568161114/ > /dev/null 2>&1 && sleep 0' 13118 1727204096.27212: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204096.27281: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.27303: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.27362: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.27629: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.27731: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204096.27829: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.28277: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204096.28325: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204096.28359: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204096.28384: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.28416: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.28446: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.28499: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.28519: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204096.28543: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.28711: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.28841: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204096.28870: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.29058: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.30870: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204096.30874: stdout chunk (state=3): >>><<< 13118 1727204096.30876: stderr chunk (state=3): >>><<< 13118 1727204096.31257: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204096.31261: handler run complete 13118 1727204096.31265: attempt loop complete, returning result 13118 1727204096.31268: _execute() done 13118 1727204096.31270: dumping result to json 13118 1727204096.31272: done dumping result, returning 13118 1727204096.31276: done running TaskExecutor() for managed-node2/TASK: fedora.linux_system_roles.network : Re-test connectivity [0affcd87-79f5-56a3-0a64-000000000091] 13118 1727204096.31297: sending task result for task 0affcd87-79f5-56a3-0a64-000000000091 ok: [managed-node2] => { "changed": false, "ping": "pong" } 13118 1727204096.32006: no more pending results, returning what we have 13118 1727204096.32010: results queue empty 13118 1727204096.32011: checking for any_errors_fatal 13118 1727204096.32016: done checking for any_errors_fatal 13118 1727204096.32017: checking for max_fail_percentage 13118 1727204096.32018: done checking for max_fail_percentage 13118 1727204096.32019: checking to see if all hosts have failed and the running result is not ok 13118 1727204096.32020: done checking to see if all hosts have failed 13118 1727204096.32021: getting the remaining hosts for this loop 13118 1727204096.32022: done getting the remaining hosts for this loop 13118 1727204096.32025: getting the next task for host managed-node2 13118 1727204096.32036: done getting next task for host managed-node2 13118 1727204096.32038: ^ task is: TASK: meta (role_complete) 13118 1727204096.32042: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204096.32052: getting variables 13118 1727204096.32054: in VariableManager get_vars() 13118 1727204096.32108: Calling all_inventory to load vars for managed-node2 13118 1727204096.32115: Calling groups_inventory to load vars for managed-node2 13118 1727204096.32118: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204096.32128: Calling all_plugins_play to load vars for managed-node2 13118 1727204096.32133: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204096.32136: Calling groups_plugins_play to load vars for managed-node2 13118 1727204096.38203: done sending task result for task 0affcd87-79f5-56a3-0a64-000000000091 13118 1727204096.38209: WORKER PROCESS EXITING 13118 1727204096.39813: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204096.41693: done with get_vars() 13118 1727204096.41735: done getting variables 13118 1727204096.41827: done queuing things up, now waiting for results queue to drain 13118 1727204096.41832: results queue empty 13118 1727204096.41833: checking for any_errors_fatal 13118 1727204096.41837: done checking for any_errors_fatal 13118 1727204096.41838: checking for max_fail_percentage 13118 1727204096.41839: done checking for max_fail_percentage 13118 1727204096.41840: checking to see if all hosts have failed and the running result is not ok 13118 1727204096.41840: done checking to see if all hosts have failed 13118 1727204096.41841: getting the remaining hosts for this loop 13118 1727204096.41842: done getting the remaining hosts for this loop 13118 1727204096.41845: getting the next task for host managed-node2 13118 1727204096.41854: done getting next task for host managed-node2 13118 1727204096.41856: ^ task is: TASK: Delete the device '{{ controller_device }}' 13118 1727204096.41858: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204096.41861: getting variables 13118 1727204096.41862: in VariableManager get_vars() 13118 1727204096.41879: Calling all_inventory to load vars for managed-node2 13118 1727204096.41882: Calling groups_inventory to load vars for managed-node2 13118 1727204096.41884: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204096.41889: Calling all_plugins_play to load vars for managed-node2 13118 1727204096.41891: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204096.41894: Calling groups_plugins_play to load vars for managed-node2 13118 1727204096.43395: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204096.45169: done with get_vars() 13118 1727204096.45193: done getting variables 13118 1727204096.45225: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) 13118 1727204096.45305: variable 'controller_device' from source: play vars TASK [Delete the device 'nm-bond'] ********************************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml:114 Tuesday 24 September 2024 14:54:56 -0400 (0:00:00.519) 0:00:33.753 ***** 13118 1727204096.45329: entering _queue_task() for managed-node2/command 13118 1727204096.45583: worker is 1 (out of 1 available) 13118 1727204096.45597: exiting _queue_task() for managed-node2/command 13118 1727204096.45611: done queuing things up, now waiting for results queue to drain 13118 1727204096.45613: waiting for pending results... 13118 1727204096.45800: running TaskExecutor() for managed-node2/TASK: Delete the device 'nm-bond' 13118 1727204096.45879: in run() - task 0affcd87-79f5-56a3-0a64-0000000000c1 13118 1727204096.45891: variable 'ansible_search_path' from source: unknown 13118 1727204096.45921: calling self._execute() 13118 1727204096.46002: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204096.46005: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204096.46015: variable 'omit' from source: magic vars 13118 1727204096.46286: variable 'ansible_distribution_major_version' from source: facts 13118 1727204096.46297: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204096.46303: variable 'omit' from source: magic vars 13118 1727204096.46319: variable 'omit' from source: magic vars 13118 1727204096.46394: variable 'controller_device' from source: play vars 13118 1727204096.46404: variable 'omit' from source: magic vars 13118 1727204096.46439: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204096.46470: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204096.46487: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204096.46503: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204096.46512: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204096.46537: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204096.46540: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204096.46543: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204096.46615: Set connection var ansible_timeout to 10 13118 1727204096.46622: Set connection var ansible_pipelining to False 13118 1727204096.46625: Set connection var ansible_connection to ssh 13118 1727204096.46632: Set connection var ansible_shell_executable to /bin/sh 13118 1727204096.46635: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204096.46637: Set connection var ansible_shell_type to sh 13118 1727204096.46656: variable 'ansible_shell_executable' from source: unknown 13118 1727204096.46659: variable 'ansible_connection' from source: unknown 13118 1727204096.46661: variable 'ansible_module_compression' from source: unknown 13118 1727204096.46665: variable 'ansible_shell_type' from source: unknown 13118 1727204096.46667: variable 'ansible_shell_executable' from source: unknown 13118 1727204096.46670: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204096.46672: variable 'ansible_pipelining' from source: unknown 13118 1727204096.46676: variable 'ansible_timeout' from source: unknown 13118 1727204096.46679: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204096.46779: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204096.46789: variable 'omit' from source: magic vars 13118 1727204096.46795: starting attempt loop 13118 1727204096.46798: running the handler 13118 1727204096.46810: _low_level_execute_command(): starting 13118 1727204096.46819: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204096.47378: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.47483: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.47488: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.47636: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.47639: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204096.47641: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.47688: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.49322: stdout chunk (state=3): >>>/root <<< 13118 1727204096.49450: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204096.49500: stderr chunk (state=3): >>><<< 13118 1727204096.49506: stdout chunk (state=3): >>><<< 13118 1727204096.49533: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204096.49555: _low_level_execute_command(): starting 13118 1727204096.49562: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204096.4953952-15810-29304200035065 `" && echo ansible-tmp-1727204096.4953952-15810-29304200035065="` echo /root/.ansible/tmp/ansible-tmp-1727204096.4953952-15810-29304200035065 `" ) && sleep 0' 13118 1727204096.50025: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.50038: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.50065: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.50083: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.50133: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.50142: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204096.50151: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.50200: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.52074: stdout chunk (state=3): >>>ansible-tmp-1727204096.4953952-15810-29304200035065=/root/.ansible/tmp/ansible-tmp-1727204096.4953952-15810-29304200035065 <<< 13118 1727204096.52199: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204096.52296: stderr chunk (state=3): >>><<< 13118 1727204096.52299: stdout chunk (state=3): >>><<< 13118 1727204096.52320: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204096.4953952-15810-29304200035065=/root/.ansible/tmp/ansible-tmp-1727204096.4953952-15810-29304200035065 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204096.52353: variable 'ansible_module_compression' from source: unknown 13118 1727204096.52410: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13118 1727204096.52448: variable 'ansible_facts' from source: unknown 13118 1727204096.52524: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204096.4953952-15810-29304200035065/AnsiballZ_command.py 13118 1727204096.52671: Sending initial data 13118 1727204096.52674: Sent initial data (155 bytes) 13118 1727204096.53643: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204096.53651: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.53662: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.53679: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.53718: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.53725: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204096.53739: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.53754: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204096.53761: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204096.53771: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204096.53777: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.53786: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.53797: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.53804: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.53811: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204096.53821: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.53903: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.53918: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204096.53921: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.54320: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.55763: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204096.55799: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204096.55840: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpv8qxddb3 /root/.ansible/tmp/ansible-tmp-1727204096.4953952-15810-29304200035065/AnsiballZ_command.py <<< 13118 1727204096.55890: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204096.57073: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204096.57192: stderr chunk (state=3): >>><<< 13118 1727204096.57195: stdout chunk (state=3): >>><<< 13118 1727204096.57222: done transferring module to remote 13118 1727204096.57229: _low_level_execute_command(): starting 13118 1727204096.57237: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204096.4953952-15810-29304200035065/ /root/.ansible/tmp/ansible-tmp-1727204096.4953952-15810-29304200035065/AnsiballZ_command.py && sleep 0' 13118 1727204096.58653: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204096.58893: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.58897: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.58913: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.59112: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.59124: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204096.59127: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.59147: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204096.59155: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204096.59162: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204096.59172: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.59182: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.59194: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.59203: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.59211: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204096.59219: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.59297: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.59311: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204096.59325: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.59632: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.61462: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204096.61468: stdout chunk (state=3): >>><<< 13118 1727204096.61476: stderr chunk (state=3): >>><<< 13118 1727204096.61494: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204096.61498: _low_level_execute_command(): starting 13118 1727204096.61501: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204096.4953952-15810-29304200035065/AnsiballZ_command.py && sleep 0' 13118 1727204096.62151: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204096.62155: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.62158: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.62189: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.62212: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.62219: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204096.62229: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.62246: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204096.62252: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204096.62259: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204096.62267: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.62283: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.62291: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.62298: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.62305: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204096.62313: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.62385: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.62410: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204096.62414: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.62781: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.76386: stdout chunk (state=3): >>> {"changed": true, "stdout": "", "stderr": "Cannot find device \"nm-bond\"", "rc": 1, "cmd": ["ip", "link", "del", "nm-bond"], "start": "2024-09-24 14:54:56.755647", "end": "2024-09-24 14:54:56.763059", "delta": "0:00:00.007412", "failed": true, "msg": "non-zero return code", "invocation": {"module_args": {"_raw_params": "ip link del nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13118 1727204096.77496: stderr chunk (state=3): >>>debug2: Received exit status from master 1 Shared connection to 10.31.13.78 closed. <<< 13118 1727204096.77581: stderr chunk (state=3): >>><<< 13118 1727204096.77584: stdout chunk (state=3): >>><<< 13118 1727204096.77628: _low_level_execute_command() done: rc=1, stdout= {"changed": true, "stdout": "", "stderr": "Cannot find device \"nm-bond\"", "rc": 1, "cmd": ["ip", "link", "del", "nm-bond"], "start": "2024-09-24 14:54:56.755647", "end": "2024-09-24 14:54:56.763059", "delta": "0:00:00.007412", "failed": true, "msg": "non-zero return code", "invocation": {"module_args": {"_raw_params": "ip link del nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 1 Shared connection to 10.31.13.78 closed. 13118 1727204096.77656: done with _execute_module (ansible.legacy.command, {'_raw_params': 'ip link del nm-bond', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204096.4953952-15810-29304200035065/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204096.77676: _low_level_execute_command(): starting 13118 1727204096.77681: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204096.4953952-15810-29304200035065/ > /dev/null 2>&1 && sleep 0' 13118 1727204096.78448: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.78469: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.78550: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204096.78568: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.78591: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration <<< 13118 1727204096.78603: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.78628: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204096.78633: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.78742: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204096.78746: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.78833: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.80645: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204096.80747: stderr chunk (state=3): >>><<< 13118 1727204096.80750: stdout chunk (state=3): >>><<< 13118 1727204096.80762: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204096.80770: handler run complete 13118 1727204096.80800: Evaluated conditional (False): False 13118 1727204096.80804: Evaluated conditional (False): False 13118 1727204096.80815: attempt loop complete, returning result 13118 1727204096.80818: _execute() done 13118 1727204096.80824: dumping result to json 13118 1727204096.80848: done dumping result, returning 13118 1727204096.80857: done running TaskExecutor() for managed-node2/TASK: Delete the device 'nm-bond' [0affcd87-79f5-56a3-0a64-0000000000c1] 13118 1727204096.80863: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000c1 13118 1727204096.81034: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000c1 13118 1727204096.81036: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "cmd": [ "ip", "link", "del", "nm-bond" ], "delta": "0:00:00.007412", "end": "2024-09-24 14:54:56.763059", "failed_when_result": false, "rc": 1, "start": "2024-09-24 14:54:56.755647" } STDERR: Cannot find device "nm-bond" MSG: non-zero return code 13118 1727204096.81131: no more pending results, returning what we have 13118 1727204096.81135: results queue empty 13118 1727204096.81136: checking for any_errors_fatal 13118 1727204096.81138: done checking for any_errors_fatal 13118 1727204096.81139: checking for max_fail_percentage 13118 1727204096.81141: done checking for max_fail_percentage 13118 1727204096.81141: checking to see if all hosts have failed and the running result is not ok 13118 1727204096.81142: done checking to see if all hosts have failed 13118 1727204096.81143: getting the remaining hosts for this loop 13118 1727204096.81146: done getting the remaining hosts for this loop 13118 1727204096.81149: getting the next task for host managed-node2 13118 1727204096.81163: done getting next task for host managed-node2 13118 1727204096.81176: ^ task is: TASK: Remove test interfaces 13118 1727204096.81181: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204096.81189: getting variables 13118 1727204096.81192: in VariableManager get_vars() 13118 1727204096.81244: Calling all_inventory to load vars for managed-node2 13118 1727204096.81249: Calling groups_inventory to load vars for managed-node2 13118 1727204096.81252: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204096.81265: Calling all_plugins_play to load vars for managed-node2 13118 1727204096.81268: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204096.81271: Calling groups_plugins_play to load vars for managed-node2 13118 1727204096.82699: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204096.84387: done with get_vars() 13118 1727204096.84419: done getting variables 13118 1727204096.84510: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Remove test interfaces] ************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml:3 Tuesday 24 September 2024 14:54:56 -0400 (0:00:00.392) 0:00:34.145 ***** 13118 1727204096.84556: entering _queue_task() for managed-node2/shell 13118 1727204096.84833: worker is 1 (out of 1 available) 13118 1727204096.84848: exiting _queue_task() for managed-node2/shell 13118 1727204096.84865: done queuing things up, now waiting for results queue to drain 13118 1727204096.84868: waiting for pending results... 13118 1727204096.85150: running TaskExecutor() for managed-node2/TASK: Remove test interfaces 13118 1727204096.85266: in run() - task 0affcd87-79f5-56a3-0a64-0000000000c5 13118 1727204096.85280: variable 'ansible_search_path' from source: unknown 13118 1727204096.85284: variable 'ansible_search_path' from source: unknown 13118 1727204096.85311: calling self._execute() 13118 1727204096.85397: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204096.85438: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204096.85442: variable 'omit' from source: magic vars 13118 1727204096.85733: variable 'ansible_distribution_major_version' from source: facts 13118 1727204096.85745: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204096.85751: variable 'omit' from source: magic vars 13118 1727204096.85811: variable 'omit' from source: magic vars 13118 1727204096.86001: variable 'dhcp_interface1' from source: play vars 13118 1727204096.86005: variable 'dhcp_interface2' from source: play vars 13118 1727204096.86020: variable 'omit' from source: magic vars 13118 1727204096.86056: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204096.86089: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204096.86109: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204096.86122: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204096.86131: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204096.86159: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204096.86162: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204096.86166: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204096.86234: Set connection var ansible_timeout to 10 13118 1727204096.86245: Set connection var ansible_pipelining to False 13118 1727204096.86248: Set connection var ansible_connection to ssh 13118 1727204096.86257: Set connection var ansible_shell_executable to /bin/sh 13118 1727204096.86260: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204096.86263: Set connection var ansible_shell_type to sh 13118 1727204096.86280: variable 'ansible_shell_executable' from source: unknown 13118 1727204096.86283: variable 'ansible_connection' from source: unknown 13118 1727204096.86285: variable 'ansible_module_compression' from source: unknown 13118 1727204096.86288: variable 'ansible_shell_type' from source: unknown 13118 1727204096.86290: variable 'ansible_shell_executable' from source: unknown 13118 1727204096.86292: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204096.86295: variable 'ansible_pipelining' from source: unknown 13118 1727204096.86297: variable 'ansible_timeout' from source: unknown 13118 1727204096.86302: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204096.86405: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204096.86417: variable 'omit' from source: magic vars 13118 1727204096.86420: starting attempt loop 13118 1727204096.86423: running the handler 13118 1727204096.86433: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204096.86450: _low_level_execute_command(): starting 13118 1727204096.86457: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204096.86994: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.87013: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.87032: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found <<< 13118 1727204096.87045: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.87057: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.87109: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.87116: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204096.87128: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.87184: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.88770: stdout chunk (state=3): >>>/root <<< 13118 1727204096.88976: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204096.88980: stderr chunk (state=3): >>><<< 13118 1727204096.88983: stdout chunk (state=3): >>><<< 13118 1727204096.89011: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204096.89024: _low_level_execute_command(): starting 13118 1727204096.89033: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204096.8901007-15830-98420178943630 `" && echo ansible-tmp-1727204096.8901007-15830-98420178943630="` echo /root/.ansible/tmp/ansible-tmp-1727204096.8901007-15830-98420178943630 `" ) && sleep 0' 13118 1727204096.89737: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204096.89763: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.89768: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.89825: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.89828: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.89831: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204096.89864: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.89867: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204096.89875: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204096.89882: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204096.89910: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.89915: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.89918: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.89921: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.89928: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204096.89941: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.90024: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.90058: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204096.90061: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.90153: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.92009: stdout chunk (state=3): >>>ansible-tmp-1727204096.8901007-15830-98420178943630=/root/.ansible/tmp/ansible-tmp-1727204096.8901007-15830-98420178943630 <<< 13118 1727204096.92121: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204096.92186: stderr chunk (state=3): >>><<< 13118 1727204096.92190: stdout chunk (state=3): >>><<< 13118 1727204096.92212: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204096.8901007-15830-98420178943630=/root/.ansible/tmp/ansible-tmp-1727204096.8901007-15830-98420178943630 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204096.92246: variable 'ansible_module_compression' from source: unknown 13118 1727204096.92292: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13118 1727204096.92346: variable 'ansible_facts' from source: unknown 13118 1727204096.92423: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204096.8901007-15830-98420178943630/AnsiballZ_command.py 13118 1727204096.92628: Sending initial data 13118 1727204096.92631: Sent initial data (155 bytes) 13118 1727204096.94554: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204096.94563: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.94576: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.94590: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.94630: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.94639: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204096.94650: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.94662: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204096.94673: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204096.94680: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204096.94687: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.94697: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.94708: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.94718: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.94721: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204096.94733: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.94807: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.94826: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204096.94842: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.94909: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204096.96619: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204096.96662: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204096.96701: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpncok1kbv /root/.ansible/tmp/ansible-tmp-1727204096.8901007-15830-98420178943630/AnsiballZ_command.py <<< 13118 1727204096.96730: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204096.97867: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204096.97962: stderr chunk (state=3): >>><<< 13118 1727204096.97967: stdout chunk (state=3): >>><<< 13118 1727204096.97990: done transferring module to remote 13118 1727204096.98001: _low_level_execute_command(): starting 13118 1727204096.98021: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204096.8901007-15830-98420178943630/ /root/.ansible/tmp/ansible-tmp-1727204096.8901007-15830-98420178943630/AnsiballZ_command.py && sleep 0' 13118 1727204096.98818: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204096.98827: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.98837: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.98858: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.98905: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.98912: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204096.98923: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.98937: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204096.98944: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204096.98953: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204096.98964: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204096.98981: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204096.98992: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204096.99000: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204096.99007: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204096.99016: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204096.99101: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204096.99119: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204096.99134: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204096.99199: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204097.00987: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204097.00991: stdout chunk (state=3): >>><<< 13118 1727204097.00993: stderr chunk (state=3): >>><<< 13118 1727204097.01013: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204097.01017: _low_level_execute_command(): starting 13118 1727204097.01019: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204096.8901007-15830-98420178943630/AnsiballZ_command.py && sleep 0' 13118 1727204097.01737: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204097.01746: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.01762: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.01784: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.01838: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.01842: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204097.01844: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.01859: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204097.01867: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204097.01876: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204097.01884: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.01893: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.02033: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.02048: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.02052: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204097.02090: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.02093: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204097.02096: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204097.02102: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204097.02143: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204097.21161: stdout chunk (state=3): >>> {"changed": true, "stdout": "", "stderr": "+ exec\n+ rc=0\n+ ip link delete test1\n+ '[' 0 '!=' 0 ']'\n+ ip link delete test2\n+ '[' 0 '!=' 0 ']'\n+ ip link delete testbr\n+ '[' 0 '!=' 0 ']'", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "start": "2024-09-24 14:54:57.154865", "end": "2024-09-24 14:54:57.209159", "delta": "0:00:00.054294", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13118 1727204097.22867: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204097.22872: stderr chunk (state=3): >>><<< 13118 1727204097.22874: stdout chunk (state=3): >>><<< 13118 1727204097.22877: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "+ exec\n+ rc=0\n+ ip link delete test1\n+ '[' 0 '!=' 0 ']'\n+ ip link delete test2\n+ '[' 0 '!=' 0 ']'\n+ ip link delete testbr\n+ '[' 0 '!=' 0 ']'", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "start": "2024-09-24 14:54:57.154865", "end": "2024-09-24 14:54:57.209159", "delta": "0:00:00.054294", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204097.22884: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc="$?"\nif [ "$rc" != 0 ]; then\n echo ERROR - could not delete link test1 - error "$rc"\nfi\nip link delete test2 || rc="$?"\nif [ "$rc" != 0 ]; then\n echo ERROR - could not delete link test2 - error "$rc"\nfi\nip link delete testbr || rc="$?"\nif [ "$rc" != 0 ]; then\n echo ERROR - could not delete link testbr - error "$rc"\nfi\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204096.8901007-15830-98420178943630/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204097.22887: _low_level_execute_command(): starting 13118 1727204097.22889: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204096.8901007-15830-98420178943630/ > /dev/null 2>&1 && sleep 0' 13118 1727204097.23920: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204097.23942: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.23960: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.23983: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.24042: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.24054: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204097.24080: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.24103: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204097.24119: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204097.24129: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204097.24145: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.24158: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.24177: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.24189: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.24201: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204097.24225: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.24304: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204097.24339: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204097.24356: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204097.24443: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204097.26286: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204097.26406: stderr chunk (state=3): >>><<< 13118 1727204097.26411: stdout chunk (state=3): >>><<< 13118 1727204097.26470: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204097.26474: handler run complete 13118 1727204097.26477: Evaluated conditional (False): False 13118 1727204097.26676: attempt loop complete, returning result 13118 1727204097.26679: _execute() done 13118 1727204097.26682: dumping result to json 13118 1727204097.26684: done dumping result, returning 13118 1727204097.26686: done running TaskExecutor() for managed-node2/TASK: Remove test interfaces [0affcd87-79f5-56a3-0a64-0000000000c5] 13118 1727204097.26688: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000c5 13118 1727204097.26771: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000c5 13118 1727204097.26776: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "cmd": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "delta": "0:00:00.054294", "end": "2024-09-24 14:54:57.209159", "rc": 0, "start": "2024-09-24 14:54:57.154865" } STDERR: + exec + rc=0 + ip link delete test1 + '[' 0 '!=' 0 ']' + ip link delete test2 + '[' 0 '!=' 0 ']' + ip link delete testbr + '[' 0 '!=' 0 ']' 13118 1727204097.26864: no more pending results, returning what we have 13118 1727204097.26875: results queue empty 13118 1727204097.26875: checking for any_errors_fatal 13118 1727204097.26886: done checking for any_errors_fatal 13118 1727204097.26887: checking for max_fail_percentage 13118 1727204097.26889: done checking for max_fail_percentage 13118 1727204097.26890: checking to see if all hosts have failed and the running result is not ok 13118 1727204097.26890: done checking to see if all hosts have failed 13118 1727204097.26891: getting the remaining hosts for this loop 13118 1727204097.26892: done getting the remaining hosts for this loop 13118 1727204097.26896: getting the next task for host managed-node2 13118 1727204097.26905: done getting next task for host managed-node2 13118 1727204097.26907: ^ task is: TASK: Stop dnsmasq/radvd services 13118 1727204097.26911: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204097.26920: getting variables 13118 1727204097.26922: in VariableManager get_vars() 13118 1727204097.26971: Calling all_inventory to load vars for managed-node2 13118 1727204097.26974: Calling groups_inventory to load vars for managed-node2 13118 1727204097.26978: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204097.26991: Calling all_plugins_play to load vars for managed-node2 13118 1727204097.26994: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204097.26998: Calling groups_plugins_play to load vars for managed-node2 13118 1727204097.30107: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204097.31870: done with get_vars() 13118 1727204097.31910: done getting variables 13118 1727204097.31983: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Stop dnsmasq/radvd services] ********************************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml:23 Tuesday 24 September 2024 14:54:57 -0400 (0:00:00.474) 0:00:34.620 ***** 13118 1727204097.32026: entering _queue_task() for managed-node2/shell 13118 1727204097.32402: worker is 1 (out of 1 available) 13118 1727204097.32416: exiting _queue_task() for managed-node2/shell 13118 1727204097.32434: done queuing things up, now waiting for results queue to drain 13118 1727204097.32436: waiting for pending results... 13118 1727204097.32747: running TaskExecutor() for managed-node2/TASK: Stop dnsmasq/radvd services 13118 1727204097.32912: in run() - task 0affcd87-79f5-56a3-0a64-0000000000c6 13118 1727204097.32937: variable 'ansible_search_path' from source: unknown 13118 1727204097.32945: variable 'ansible_search_path' from source: unknown 13118 1727204097.32995: calling self._execute() 13118 1727204097.33105: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204097.33116: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204097.33129: variable 'omit' from source: magic vars 13118 1727204097.33532: variable 'ansible_distribution_major_version' from source: facts 13118 1727204097.33551: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204097.33562: variable 'omit' from source: magic vars 13118 1727204097.33616: variable 'omit' from source: magic vars 13118 1727204097.33665: variable 'omit' from source: magic vars 13118 1727204097.33708: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204097.33767: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204097.33791: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204097.33811: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204097.33826: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204097.33869: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204097.33877: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204097.33883: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204097.34003: Set connection var ansible_timeout to 10 13118 1727204097.34023: Set connection var ansible_pipelining to False 13118 1727204097.34033: Set connection var ansible_connection to ssh 13118 1727204097.34045: Set connection var ansible_shell_executable to /bin/sh 13118 1727204097.34060: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204097.34070: Set connection var ansible_shell_type to sh 13118 1727204097.34109: variable 'ansible_shell_executable' from source: unknown 13118 1727204097.34117: variable 'ansible_connection' from source: unknown 13118 1727204097.34124: variable 'ansible_module_compression' from source: unknown 13118 1727204097.34134: variable 'ansible_shell_type' from source: unknown 13118 1727204097.34141: variable 'ansible_shell_executable' from source: unknown 13118 1727204097.34147: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204097.34154: variable 'ansible_pipelining' from source: unknown 13118 1727204097.34162: variable 'ansible_timeout' from source: unknown 13118 1727204097.34178: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204097.34335: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204097.34353: variable 'omit' from source: magic vars 13118 1727204097.34366: starting attempt loop 13118 1727204097.34375: running the handler 13118 1727204097.34396: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204097.34423: _low_level_execute_command(): starting 13118 1727204097.34439: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204097.35524: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204097.35545: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.35560: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.35589: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.35639: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.35653: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204097.35670: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.35690: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204097.35707: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204097.35719: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204097.35732: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.35746: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.35760: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.35774: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.35785: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204097.35800: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.35885: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204097.35912: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204097.35940: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204097.36024: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204097.37710: stdout chunk (state=3): >>>/root <<< 13118 1727204097.37887: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204097.37891: stdout chunk (state=3): >>><<< 13118 1727204097.37902: stderr chunk (state=3): >>><<< 13118 1727204097.37927: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204097.37942: _low_level_execute_command(): starting 13118 1727204097.37956: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204097.3792624-15855-238712167745727 `" && echo ansible-tmp-1727204097.3792624-15855-238712167745727="` echo /root/.ansible/tmp/ansible-tmp-1727204097.3792624-15855-238712167745727 `" ) && sleep 0' 13118 1727204097.38613: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204097.38622: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.38636: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.38648: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.38694: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.38697: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204097.38709: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.38722: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204097.38729: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204097.38735: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204097.38743: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.38752: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.38763: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.38772: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.38781: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204097.38788: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.38861: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204097.38882: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204097.38894: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204097.38967: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204097.40886: stdout chunk (state=3): >>>ansible-tmp-1727204097.3792624-15855-238712167745727=/root/.ansible/tmp/ansible-tmp-1727204097.3792624-15855-238712167745727 <<< 13118 1727204097.41073: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204097.41077: stderr chunk (state=3): >>><<< 13118 1727204097.41080: stdout chunk (state=3): >>><<< 13118 1727204097.41115: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204097.3792624-15855-238712167745727=/root/.ansible/tmp/ansible-tmp-1727204097.3792624-15855-238712167745727 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204097.41148: variable 'ansible_module_compression' from source: unknown 13118 1727204097.41207: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13118 1727204097.41249: variable 'ansible_facts' from source: unknown 13118 1727204097.41323: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204097.3792624-15855-238712167745727/AnsiballZ_command.py 13118 1727204097.41482: Sending initial data 13118 1727204097.41485: Sent initial data (156 bytes) 13118 1727204097.42517: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204097.42527: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.42537: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.42552: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.42601: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.42610: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204097.42619: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.42635: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204097.42641: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204097.42648: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204097.42656: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.42667: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.42679: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.42688: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.42696: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204097.42710: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.42782: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204097.42801: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204097.42816: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204097.42882: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204097.44692: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204097.44835: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204097.44840: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpir049hk3 /root/.ansible/tmp/ansible-tmp-1727204097.3792624-15855-238712167745727/AnsiballZ_command.py <<< 13118 1727204097.44843: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204097.46452: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204097.46589: stderr chunk (state=3): >>><<< 13118 1727204097.46592: stdout chunk (state=3): >>><<< 13118 1727204097.46595: done transferring module to remote 13118 1727204097.46597: _low_level_execute_command(): starting 13118 1727204097.46600: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204097.3792624-15855-238712167745727/ /root/.ansible/tmp/ansible-tmp-1727204097.3792624-15855-238712167745727/AnsiballZ_command.py && sleep 0' 13118 1727204097.47550: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.47554: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.47578: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.47623: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204097.47641: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.47659: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204097.47675: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204097.47688: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204097.47700: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.47714: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.47729: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.47745: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.47758: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204097.47782: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.47857: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204097.47881: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204097.47898: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204097.47971: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204097.49871: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204097.49875: stdout chunk (state=3): >>><<< 13118 1727204097.49900: stderr chunk (state=3): >>><<< 13118 1727204097.49973: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204097.49977: _low_level_execute_command(): starting 13118 1727204097.49979: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204097.3792624-15855-238712167745727/AnsiballZ_command.py && sleep 0' 13118 1727204097.51656: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204097.51675: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.51691: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.51708: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.51754: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.51768: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204097.51782: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.51799: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204097.51809: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204097.51819: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204097.51832: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.51847: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.51865: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.51880: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.51892: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204097.51906: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.51986: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204097.52189: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204097.52205: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204097.52295: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204097.67858: stdout chunk (state=3): >>> {"changed": true, "stdout": "", "stderr": "+ exec\n+ pkill -F /run/dhcp_testbr.pid\n+ rm -rf /run/dhcp_testbr.pid\n+ rm -rf /run/dhcp_testbr.lease\n+ grep 'release 6' /etc/redhat-release\n+ systemctl is-active firewalld\ninactive", "rc": 0, "cmd": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "start": "2024-09-24 14:54:57.655762", "end": "2024-09-24 14:54:57.677247", "delta": "0:00:00.021485", "msg": "", "invocation": {"module_args": {"_raw_params": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13118 1727204097.69259: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204097.69263: stdout chunk (state=3): >>><<< 13118 1727204097.69268: stderr chunk (state=3): >>><<< 13118 1727204097.69434: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "+ exec\n+ pkill -F /run/dhcp_testbr.pid\n+ rm -rf /run/dhcp_testbr.pid\n+ rm -rf /run/dhcp_testbr.lease\n+ grep 'release 6' /etc/redhat-release\n+ systemctl is-active firewalld\ninactive", "rc": 0, "cmd": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "start": "2024-09-24 14:54:57.655762", "end": "2024-09-24 14:54:57.677247", "delta": "0:00:00.021485", "msg": "", "invocation": {"module_args": {"_raw_params": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204097.69438: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep \'release 6\' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service="$service"; then\n firewall-cmd --remove-service "$service"\n fi\n done\nfi\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204097.3792624-15855-238712167745727/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204097.69441: _low_level_execute_command(): starting 13118 1727204097.69444: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204097.3792624-15855-238712167745727/ > /dev/null 2>&1 && sleep 0' 13118 1727204097.71456: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204097.71466: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.71476: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.71489: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.71537: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.71544: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204097.71556: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.71570: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204097.71576: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204097.71582: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204097.71589: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204097.71597: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204097.71607: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204097.71616: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204097.71628: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204097.71639: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204097.71713: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204097.71724: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204097.71739: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204097.71817: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204097.73770: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204097.73774: stdout chunk (state=3): >>><<< 13118 1727204097.73776: stderr chunk (state=3): >>><<< 13118 1727204097.74245: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204097.74252: handler run complete 13118 1727204097.74254: Evaluated conditional (False): False 13118 1727204097.74256: attempt loop complete, returning result 13118 1727204097.74258: _execute() done 13118 1727204097.74259: dumping result to json 13118 1727204097.74261: done dumping result, returning 13118 1727204097.74263: done running TaskExecutor() for managed-node2/TASK: Stop dnsmasq/radvd services [0affcd87-79f5-56a3-0a64-0000000000c6] 13118 1727204097.74266: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000c6 13118 1727204097.74346: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000c6 13118 1727204097.74350: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "cmd": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "delta": "0:00:00.021485", "end": "2024-09-24 14:54:57.677247", "rc": 0, "start": "2024-09-24 14:54:57.655762" } STDERR: + exec + pkill -F /run/dhcp_testbr.pid + rm -rf /run/dhcp_testbr.pid + rm -rf /run/dhcp_testbr.lease + grep 'release 6' /etc/redhat-release + systemctl is-active firewalld inactive 13118 1727204097.74419: no more pending results, returning what we have 13118 1727204097.74423: results queue empty 13118 1727204097.74424: checking for any_errors_fatal 13118 1727204097.74430: done checking for any_errors_fatal 13118 1727204097.74431: checking for max_fail_percentage 13118 1727204097.74433: done checking for max_fail_percentage 13118 1727204097.74434: checking to see if all hosts have failed and the running result is not ok 13118 1727204097.74434: done checking to see if all hosts have failed 13118 1727204097.74435: getting the remaining hosts for this loop 13118 1727204097.74436: done getting the remaining hosts for this loop 13118 1727204097.74440: getting the next task for host managed-node2 13118 1727204097.74448: done getting next task for host managed-node2 13118 1727204097.74453: ^ task is: TASK: Restore the /etc/resolv.conf for initscript 13118 1727204097.74456: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204097.74459: getting variables 13118 1727204097.74460: in VariableManager get_vars() 13118 1727204097.74497: Calling all_inventory to load vars for managed-node2 13118 1727204097.74499: Calling groups_inventory to load vars for managed-node2 13118 1727204097.74501: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204097.74510: Calling all_plugins_play to load vars for managed-node2 13118 1727204097.74513: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204097.74515: Calling groups_plugins_play to load vars for managed-node2 13118 1727204097.77055: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204097.80305: done with get_vars() 13118 1727204097.80330: done getting variables 13118 1727204097.80402: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Restore the /etc/resolv.conf for initscript] ***************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml:120 Tuesday 24 September 2024 14:54:57 -0400 (0:00:00.484) 0:00:35.104 ***** 13118 1727204097.80435: entering _queue_task() for managed-node2/command 13118 1727204097.80791: worker is 1 (out of 1 available) 13118 1727204097.80803: exiting _queue_task() for managed-node2/command 13118 1727204097.80818: done queuing things up, now waiting for results queue to drain 13118 1727204097.80819: waiting for pending results... 13118 1727204097.81144: running TaskExecutor() for managed-node2/TASK: Restore the /etc/resolv.conf for initscript 13118 1727204097.81257: in run() - task 0affcd87-79f5-56a3-0a64-0000000000c7 13118 1727204097.81278: variable 'ansible_search_path' from source: unknown 13118 1727204097.81336: calling self._execute() 13118 1727204097.81458: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204097.81462: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204097.81510: variable 'omit' from source: magic vars 13118 1727204097.82249: variable 'ansible_distribution_major_version' from source: facts 13118 1727204097.82261: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204097.82589: variable 'network_provider' from source: set_fact 13118 1727204097.82593: Evaluated conditional (network_provider == "initscripts"): False 13118 1727204097.82604: when evaluation is False, skipping this task 13118 1727204097.82715: _execute() done 13118 1727204097.82725: dumping result to json 13118 1727204097.82728: done dumping result, returning 13118 1727204097.82736: done running TaskExecutor() for managed-node2/TASK: Restore the /etc/resolv.conf for initscript [0affcd87-79f5-56a3-0a64-0000000000c7] 13118 1727204097.82742: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000c7 13118 1727204097.82851: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000c7 13118 1727204097.82855: WORKER PROCESS EXITING skipping: [managed-node2] => { "changed": false, "false_condition": "network_provider == \"initscripts\"", "skip_reason": "Conditional result was False" } 13118 1727204097.82908: no more pending results, returning what we have 13118 1727204097.82913: results queue empty 13118 1727204097.82914: checking for any_errors_fatal 13118 1727204097.82928: done checking for any_errors_fatal 13118 1727204097.82928: checking for max_fail_percentage 13118 1727204097.82932: done checking for max_fail_percentage 13118 1727204097.82933: checking to see if all hosts have failed and the running result is not ok 13118 1727204097.82934: done checking to see if all hosts have failed 13118 1727204097.82935: getting the remaining hosts for this loop 13118 1727204097.82936: done getting the remaining hosts for this loop 13118 1727204097.82941: getting the next task for host managed-node2 13118 1727204097.82951: done getting next task for host managed-node2 13118 1727204097.82954: ^ task is: TASK: Verify network state restored to default 13118 1727204097.82957: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204097.82961: getting variables 13118 1727204097.82963: in VariableManager get_vars() 13118 1727204097.83010: Calling all_inventory to load vars for managed-node2 13118 1727204097.83014: Calling groups_inventory to load vars for managed-node2 13118 1727204097.83016: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204097.83030: Calling all_plugins_play to load vars for managed-node2 13118 1727204097.83034: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204097.83037: Calling groups_plugins_play to load vars for managed-node2 13118 1727204097.85321: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204097.87989: done with get_vars() 13118 1727204097.88024: done getting variables TASK [Verify network state restored to default] ******************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml:125 Tuesday 24 September 2024 14:54:57 -0400 (0:00:00.079) 0:00:35.183 ***** 13118 1727204097.88344: entering _queue_task() for managed-node2/include_tasks 13118 1727204097.88794: worker is 1 (out of 1 available) 13118 1727204097.88806: exiting _queue_task() for managed-node2/include_tasks 13118 1727204097.88819: done queuing things up, now waiting for results queue to drain 13118 1727204097.88820: waiting for pending results... 13118 1727204097.89866: running TaskExecutor() for managed-node2/TASK: Verify network state restored to default 13118 1727204097.90185: in run() - task 0affcd87-79f5-56a3-0a64-0000000000c8 13118 1727204097.90199: variable 'ansible_search_path' from source: unknown 13118 1727204097.90241: calling self._execute() 13118 1727204097.90383: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204097.90390: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204097.90405: variable 'omit' from source: magic vars 13118 1727204097.90824: variable 'ansible_distribution_major_version' from source: facts 13118 1727204097.90839: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204097.90846: _execute() done 13118 1727204097.90850: dumping result to json 13118 1727204097.90852: done dumping result, returning 13118 1727204097.90858: done running TaskExecutor() for managed-node2/TASK: Verify network state restored to default [0affcd87-79f5-56a3-0a64-0000000000c8] 13118 1727204097.90867: sending task result for task 0affcd87-79f5-56a3-0a64-0000000000c8 13118 1727204097.90971: done sending task result for task 0affcd87-79f5-56a3-0a64-0000000000c8 13118 1727204097.90975: WORKER PROCESS EXITING 13118 1727204097.91004: no more pending results, returning what we have 13118 1727204097.91011: in VariableManager get_vars() 13118 1727204097.91069: Calling all_inventory to load vars for managed-node2 13118 1727204097.91072: Calling groups_inventory to load vars for managed-node2 13118 1727204097.91075: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204097.91090: Calling all_plugins_play to load vars for managed-node2 13118 1727204097.91094: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204097.91097: Calling groups_plugins_play to load vars for managed-node2 13118 1727204097.93251: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204097.96458: done with get_vars() 13118 1727204097.96502: variable 'ansible_search_path' from source: unknown 13118 1727204097.96520: we have included files to process 13118 1727204097.96522: generating all_blocks data 13118 1727204097.96525: done generating all_blocks data 13118 1727204097.96534: processing included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml 13118 1727204097.96536: loading included file: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml 13118 1727204097.96539: Loading data from /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml 13118 1727204097.97025: done processing included file 13118 1727204097.97028: iterating over new_blocks loaded from include file 13118 1727204097.97035: in VariableManager get_vars() 13118 1727204097.97057: done with get_vars() 13118 1727204097.97059: filtering new block on tags 13118 1727204097.97099: done filtering new block on tags 13118 1727204097.97102: done iterating over new_blocks loaded from include file included: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml for managed-node2 13118 1727204097.97109: extending task lists for all hosts with included blocks 13118 1727204097.98884: done extending task lists 13118 1727204097.98886: done processing included files 13118 1727204097.98887: results queue empty 13118 1727204097.98888: checking for any_errors_fatal 13118 1727204097.98891: done checking for any_errors_fatal 13118 1727204097.98892: checking for max_fail_percentage 13118 1727204097.98893: done checking for max_fail_percentage 13118 1727204097.98894: checking to see if all hosts have failed and the running result is not ok 13118 1727204097.98895: done checking to see if all hosts have failed 13118 1727204097.98896: getting the remaining hosts for this loop 13118 1727204097.98897: done getting the remaining hosts for this loop 13118 1727204097.98899: getting the next task for host managed-node2 13118 1727204097.98904: done getting next task for host managed-node2 13118 1727204097.98906: ^ task is: TASK: Check routes and DNS 13118 1727204097.98908: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204097.98912: getting variables 13118 1727204097.98913: in VariableManager get_vars() 13118 1727204097.98934: Calling all_inventory to load vars for managed-node2 13118 1727204097.98937: Calling groups_inventory to load vars for managed-node2 13118 1727204097.98939: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204097.99062: Calling all_plugins_play to load vars for managed-node2 13118 1727204097.99069: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204097.99073: Calling groups_plugins_play to load vars for managed-node2 13118 1727204098.01098: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204098.02965: done with get_vars() 13118 1727204098.03005: done getting variables 13118 1727204098.03059: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Check routes and DNS] **************************************************** task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml:6 Tuesday 24 September 2024 14:54:58 -0400 (0:00:00.147) 0:00:35.331 ***** 13118 1727204098.03102: entering _queue_task() for managed-node2/shell 13118 1727204098.03475: worker is 1 (out of 1 available) 13118 1727204098.03487: exiting _queue_task() for managed-node2/shell 13118 1727204098.03498: done queuing things up, now waiting for results queue to drain 13118 1727204098.03504: waiting for pending results... 13118 1727204098.03810: running TaskExecutor() for managed-node2/TASK: Check routes and DNS 13118 1727204098.03930: in run() - task 0affcd87-79f5-56a3-0a64-00000000056d 13118 1727204098.03953: variable 'ansible_search_path' from source: unknown 13118 1727204098.03957: variable 'ansible_search_path' from source: unknown 13118 1727204098.03991: calling self._execute() 13118 1727204098.04095: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204098.04101: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204098.04111: variable 'omit' from source: magic vars 13118 1727204098.04521: variable 'ansible_distribution_major_version' from source: facts 13118 1727204098.04539: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204098.04545: variable 'omit' from source: magic vars 13118 1727204098.04606: variable 'omit' from source: magic vars 13118 1727204098.04645: variable 'omit' from source: magic vars 13118 1727204098.04689: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204098.04739: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204098.04757: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204098.04777: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204098.04788: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204098.04830: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204098.04833: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204098.04839: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204098.04954: Set connection var ansible_timeout to 10 13118 1727204098.04969: Set connection var ansible_pipelining to False 13118 1727204098.04972: Set connection var ansible_connection to ssh 13118 1727204098.04974: Set connection var ansible_shell_executable to /bin/sh 13118 1727204098.04981: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204098.04984: Set connection var ansible_shell_type to sh 13118 1727204098.05007: variable 'ansible_shell_executable' from source: unknown 13118 1727204098.05010: variable 'ansible_connection' from source: unknown 13118 1727204098.05013: variable 'ansible_module_compression' from source: unknown 13118 1727204098.05015: variable 'ansible_shell_type' from source: unknown 13118 1727204098.05018: variable 'ansible_shell_executable' from source: unknown 13118 1727204098.05021: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204098.05026: variable 'ansible_pipelining' from source: unknown 13118 1727204098.05039: variable 'ansible_timeout' from source: unknown 13118 1727204098.05043: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204098.05196: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204098.05205: variable 'omit' from source: magic vars 13118 1727204098.05212: starting attempt loop 13118 1727204098.05215: running the handler 13118 1727204098.05225: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204098.05259: _low_level_execute_command(): starting 13118 1727204098.05266: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204098.06114: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204098.06130: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.06146: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.06168: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.06204: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.06211: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204098.06221: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.06241: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204098.06250: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204098.06257: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204098.06272: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.06281: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.06292: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.06299: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.06306: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204098.06316: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.06391: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204098.06407: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204098.06416: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204098.06566: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204098.08193: stdout chunk (state=3): >>>/root <<< 13118 1727204098.08378: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204098.08382: stdout chunk (state=3): >>><<< 13118 1727204098.08393: stderr chunk (state=3): >>><<< 13118 1727204098.08431: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204098.08449: _low_level_execute_command(): starting 13118 1727204098.08463: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204098.084337-15901-233365317664171 `" && echo ansible-tmp-1727204098.084337-15901-233365317664171="` echo /root/.ansible/tmp/ansible-tmp-1727204098.084337-15901-233365317664171 `" ) && sleep 0' 13118 1727204098.09139: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204098.09148: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.09157: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.09172: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.09215: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.09221: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204098.09230: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.09246: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204098.09253: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204098.09259: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204098.09268: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.09277: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.09287: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.09296: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.09305: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204098.09315: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.09391: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204098.09405: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204098.09419: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204098.09814: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204098.11419: stdout chunk (state=3): >>>ansible-tmp-1727204098.084337-15901-233365317664171=/root/.ansible/tmp/ansible-tmp-1727204098.084337-15901-233365317664171 <<< 13118 1727204098.11578: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204098.11633: stderr chunk (state=3): >>><<< 13118 1727204098.11639: stdout chunk (state=3): >>><<< 13118 1727204098.11662: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204098.084337-15901-233365317664171=/root/.ansible/tmp/ansible-tmp-1727204098.084337-15901-233365317664171 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204098.11699: variable 'ansible_module_compression' from source: unknown 13118 1727204098.11752: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13118 1727204098.11791: variable 'ansible_facts' from source: unknown 13118 1727204098.11853: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204098.084337-15901-233365317664171/AnsiballZ_command.py 13118 1727204098.12010: Sending initial data 13118 1727204098.12014: Sent initial data (155 bytes) 13118 1727204098.13088: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204098.13092: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.13095: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.13097: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.13099: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.13102: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204098.13104: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.13121: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204098.13124: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204098.13126: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204098.13178: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.13181: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.13183: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.13185: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.13187: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204098.13189: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.13287: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204098.13291: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204098.13293: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204098.13565: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204098.15243: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204098.15284: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204098.15324: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpsyc866p7 /root/.ansible/tmp/ansible-tmp-1727204098.084337-15901-233365317664171/AnsiballZ_command.py <<< 13118 1727204098.15361: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204098.17071: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204098.17142: stderr chunk (state=3): >>><<< 13118 1727204098.17146: stdout chunk (state=3): >>><<< 13118 1727204098.17168: done transferring module to remote 13118 1727204098.17179: _low_level_execute_command(): starting 13118 1727204098.17184: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204098.084337-15901-233365317664171/ /root/.ansible/tmp/ansible-tmp-1727204098.084337-15901-233365317664171/AnsiballZ_command.py && sleep 0' 13118 1727204098.17920: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204098.17929: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.17939: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.17954: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.17995: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.18002: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204098.18013: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.18025: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204098.18035: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204098.18040: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204098.18048: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.18059: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.18071: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.18080: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.18087: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204098.18096: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.18158: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204098.18176: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204098.18185: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204098.18285: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204098.20082: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204098.20152: stderr chunk (state=3): >>><<< 13118 1727204098.20155: stdout chunk (state=3): >>><<< 13118 1727204098.20177: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204098.20181: _low_level_execute_command(): starting 13118 1727204098.20184: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204098.084337-15901-233365317664171/AnsiballZ_command.py && sleep 0' 13118 1727204098.21663: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204098.21882: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.21891: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.21904: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.21941: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.21949: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204098.21957: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.21971: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204098.21978: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204098.21984: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204098.21991: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.22000: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.22012: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.22019: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.22025: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204098.22035: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.22107: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204098.22190: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204098.22196: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204098.22657: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204098.37036: stdout chunk (state=3): >>> {"changed": true, "stdout": "IP\n1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000\n link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00\n inet 127.0.0.1/8 scope host lo\n valid_lft forever preferred_lft forever\n inet6 ::1/128 scope host \n valid_lft forever preferred_lft forever\n2: eth0: mtu 9001 qdisc mq state UP group default qlen 1000\n link/ether 0a:ff:ff:f5:f2:b9 brd ff:ff:ff:ff:ff:ff\n altname enX0\n inet 10.31.13.78/22 brd 10.31.15.255 scope global dynamic noprefixroute eth0\n valid_lft 3459sec preferred_lft 3459sec\n inet6 fe80::8ff:ffff:fef5:f2b9/64 scope link \n valid_lft forever preferred_lft forever\nIP ROUTE\ndefault via 10.31.12.1 dev eth0 proto dhcp src 10.31.13.78 metric 100 \n10.31.12.0/22 dev eth0 proto kernel scope link src 10.31.13.78 metric 100 \nIP -6 ROUTE\n::1 dev lo proto kernel metric 256 pref medium\nfe80::/64 dev eth0 proto kernel metric 256 pref medium\nRESOLV\n# Generated by NetworkManager\nsearch us-east-1.aws.redhat.com\nnameserver 10.29.169.13\nnameserver 10.29.170.12\nnameserver 10.2.32.1", "stderr": "", "rc": 0, "cmd": "set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n", "start": "2024-09-24 14:54:58.360336", "end": "2024-09-24 14:54:58.369372", "delta": "0:00:00.009036", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13118 1727204098.38390: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204098.38437: stderr chunk (state=3): >>><<< 13118 1727204098.38441: stdout chunk (state=3): >>><<< 13118 1727204098.38461: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "IP\n1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000\n link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00\n inet 127.0.0.1/8 scope host lo\n valid_lft forever preferred_lft forever\n inet6 ::1/128 scope host \n valid_lft forever preferred_lft forever\n2: eth0: mtu 9001 qdisc mq state UP group default qlen 1000\n link/ether 0a:ff:ff:f5:f2:b9 brd ff:ff:ff:ff:ff:ff\n altname enX0\n inet 10.31.13.78/22 brd 10.31.15.255 scope global dynamic noprefixroute eth0\n valid_lft 3459sec preferred_lft 3459sec\n inet6 fe80::8ff:ffff:fef5:f2b9/64 scope link \n valid_lft forever preferred_lft forever\nIP ROUTE\ndefault via 10.31.12.1 dev eth0 proto dhcp src 10.31.13.78 metric 100 \n10.31.12.0/22 dev eth0 proto kernel scope link src 10.31.13.78 metric 100 \nIP -6 ROUTE\n::1 dev lo proto kernel metric 256 pref medium\nfe80::/64 dev eth0 proto kernel metric 256 pref medium\nRESOLV\n# Generated by NetworkManager\nsearch us-east-1.aws.redhat.com\nnameserver 10.29.169.13\nnameserver 10.29.170.12\nnameserver 10.2.32.1", "stderr": "", "rc": 0, "cmd": "set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n", "start": "2024-09-24 14:54:58.360336", "end": "2024-09-24 14:54:58.369372", "delta": "0:00:00.009036", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204098.38512: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204098.084337-15901-233365317664171/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204098.38520: _low_level_execute_command(): starting 13118 1727204098.38526: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204098.084337-15901-233365317664171/ > /dev/null 2>&1 && sleep 0' 13118 1727204098.40271: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204098.40279: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.40289: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.40303: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.40463: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.40472: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204098.40482: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.40495: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204098.40501: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204098.40508: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204098.40515: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.40523: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.40544: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.40551: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.40559: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204098.40574: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.40645: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204098.40774: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204098.40781: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204098.40876: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204098.42857: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204098.42862: stdout chunk (state=3): >>><<< 13118 1727204098.42873: stderr chunk (state=3): >>><<< 13118 1727204098.42898: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204098.42906: handler run complete 13118 1727204098.42933: Evaluated conditional (False): False 13118 1727204098.42941: attempt loop complete, returning result 13118 1727204098.42944: _execute() done 13118 1727204098.42947: dumping result to json 13118 1727204098.42953: done dumping result, returning 13118 1727204098.42963: done running TaskExecutor() for managed-node2/TASK: Check routes and DNS [0affcd87-79f5-56a3-0a64-00000000056d] 13118 1727204098.42971: sending task result for task 0affcd87-79f5-56a3-0a64-00000000056d 13118 1727204098.43099: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000056d 13118 1727204098.43103: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "cmd": "set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n", "delta": "0:00:00.009036", "end": "2024-09-24 14:54:58.369372", "rc": 0, "start": "2024-09-24 14:54:58.360336" } STDOUT: IP 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: eth0: mtu 9001 qdisc mq state UP group default qlen 1000 link/ether 0a:ff:ff:f5:f2:b9 brd ff:ff:ff:ff:ff:ff altname enX0 inet 10.31.13.78/22 brd 10.31.15.255 scope global dynamic noprefixroute eth0 valid_lft 3459sec preferred_lft 3459sec inet6 fe80::8ff:ffff:fef5:f2b9/64 scope link valid_lft forever preferred_lft forever IP ROUTE default via 10.31.12.1 dev eth0 proto dhcp src 10.31.13.78 metric 100 10.31.12.0/22 dev eth0 proto kernel scope link src 10.31.13.78 metric 100 IP -6 ROUTE ::1 dev lo proto kernel metric 256 pref medium fe80::/64 dev eth0 proto kernel metric 256 pref medium RESOLV # Generated by NetworkManager search us-east-1.aws.redhat.com nameserver 10.29.169.13 nameserver 10.29.170.12 nameserver 10.2.32.1 13118 1727204098.43186: no more pending results, returning what we have 13118 1727204098.43191: results queue empty 13118 1727204098.43192: checking for any_errors_fatal 13118 1727204098.43194: done checking for any_errors_fatal 13118 1727204098.43195: checking for max_fail_percentage 13118 1727204098.43197: done checking for max_fail_percentage 13118 1727204098.43198: checking to see if all hosts have failed and the running result is not ok 13118 1727204098.43198: done checking to see if all hosts have failed 13118 1727204098.43199: getting the remaining hosts for this loop 13118 1727204098.43201: done getting the remaining hosts for this loop 13118 1727204098.43205: getting the next task for host managed-node2 13118 1727204098.43214: done getting next task for host managed-node2 13118 1727204098.43217: ^ task is: TASK: Verify DNS and network connectivity 13118 1727204098.43221: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 13118 1727204098.43226: getting variables 13118 1727204098.43228: in VariableManager get_vars() 13118 1727204098.43277: Calling all_inventory to load vars for managed-node2 13118 1727204098.43281: Calling groups_inventory to load vars for managed-node2 13118 1727204098.43283: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204098.43296: Calling all_plugins_play to load vars for managed-node2 13118 1727204098.43299: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204098.43302: Calling groups_plugins_play to load vars for managed-node2 13118 1727204098.47432: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204098.52292: done with get_vars() 13118 1727204098.52334: done getting variables 13118 1727204098.52557: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=True) TASK [Verify DNS and network connectivity] ************************************* task path: /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml:24 Tuesday 24 September 2024 14:54:58 -0400 (0:00:00.494) 0:00:35.825 ***** 13118 1727204098.52594: entering _queue_task() for managed-node2/shell 13118 1727204098.54340: worker is 1 (out of 1 available) 13118 1727204098.54350: exiting _queue_task() for managed-node2/shell 13118 1727204098.54361: done queuing things up, now waiting for results queue to drain 13118 1727204098.54362: waiting for pending results... 13118 1727204098.55022: running TaskExecutor() for managed-node2/TASK: Verify DNS and network connectivity 13118 1727204098.55292: in run() - task 0affcd87-79f5-56a3-0a64-00000000056e 13118 1727204098.55305: variable 'ansible_search_path' from source: unknown 13118 1727204098.55309: variable 'ansible_search_path' from source: unknown 13118 1727204098.55346: calling self._execute() 13118 1727204098.55680: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204098.55684: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204098.55818: variable 'omit' from source: magic vars 13118 1727204098.56339: variable 'ansible_distribution_major_version' from source: facts 13118 1727204098.56358: Evaluated conditional (ansible_distribution_major_version != '6'): True 13118 1727204098.56507: variable 'ansible_facts' from source: unknown 13118 1727204098.57678: Evaluated conditional (ansible_facts["distribution"] == "CentOS"): True 13118 1727204098.57683: variable 'omit' from source: magic vars 13118 1727204098.57740: variable 'omit' from source: magic vars 13118 1727204098.57780: variable 'omit' from source: magic vars 13118 1727204098.57825: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13118 1727204098.57862: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13118 1727204098.58007: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13118 1727204098.58027: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204098.58039: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13118 1727204098.58103: variable 'inventory_hostname' from source: host vars for 'managed-node2' 13118 1727204098.58106: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204098.58110: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204098.58382: Set connection var ansible_timeout to 10 13118 1727204098.58395: Set connection var ansible_pipelining to False 13118 1727204098.58398: Set connection var ansible_connection to ssh 13118 1727204098.58404: Set connection var ansible_shell_executable to /bin/sh 13118 1727204098.58409: Set connection var ansible_module_compression to ZIP_DEFLATED 13118 1727204098.58412: Set connection var ansible_shell_type to sh 13118 1727204098.58438: variable 'ansible_shell_executable' from source: unknown 13118 1727204098.58442: variable 'ansible_connection' from source: unknown 13118 1727204098.58444: variable 'ansible_module_compression' from source: unknown 13118 1727204098.58446: variable 'ansible_shell_type' from source: unknown 13118 1727204098.58449: variable 'ansible_shell_executable' from source: unknown 13118 1727204098.58451: variable 'ansible_host' from source: host vars for 'managed-node2' 13118 1727204098.58453: variable 'ansible_pipelining' from source: unknown 13118 1727204098.58457: variable 'ansible_timeout' from source: unknown 13118 1727204098.58460: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node2' 13118 1727204098.58612: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204098.58624: variable 'omit' from source: magic vars 13118 1727204098.58633: starting attempt loop 13118 1727204098.58636: running the handler 13118 1727204098.58644: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action:/usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__) (found_in_cache=True, class_only=False) 13118 1727204098.58666: _low_level_execute_command(): starting 13118 1727204098.58817: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13118 1727204098.59975: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204098.59988: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.59999: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.60014: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.60109: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.60117: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204098.60128: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.60141: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204098.60151: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204098.60161: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204098.60176: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.60185: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.60198: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.60205: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.60213: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204098.60223: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.60299: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204098.60314: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204098.60317: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204098.60504: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204098.62102: stdout chunk (state=3): >>>/root <<< 13118 1727204098.62279: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204098.62283: stdout chunk (state=3): >>><<< 13118 1727204098.62293: stderr chunk (state=3): >>><<< 13118 1727204098.62317: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204098.62333: _low_level_execute_command(): starting 13118 1727204098.62337: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727204098.6231663-15931-139618413797895 `" && echo ansible-tmp-1727204098.6231663-15931-139618413797895="` echo /root/.ansible/tmp/ansible-tmp-1727204098.6231663-15931-139618413797895 `" ) && sleep 0' 13118 1727204098.65571: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204098.65683: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.65701: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.65721: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.65772: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.65788: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204098.65803: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.65822: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204098.65837: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204098.65849: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204098.65861: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.65882: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.65902: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.65915: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.65927: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204098.65945: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.66026: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204098.66057: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204098.66084: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204098.66170: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204098.68223: stdout chunk (state=3): >>>ansible-tmp-1727204098.6231663-15931-139618413797895=/root/.ansible/tmp/ansible-tmp-1727204098.6231663-15931-139618413797895 <<< 13118 1727204098.68440: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204098.68443: stdout chunk (state=3): >>><<< 13118 1727204098.68445: stderr chunk (state=3): >>><<< 13118 1727204098.68768: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727204098.6231663-15931-139618413797895=/root/.ansible/tmp/ansible-tmp-1727204098.6231663-15931-139618413797895 , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204098.68772: variable 'ansible_module_compression' from source: unknown 13118 1727204098.68774: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13118ds5opcmo/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13118 1727204098.68776: variable 'ansible_facts' from source: unknown 13118 1727204098.68779: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727204098.6231663-15931-139618413797895/AnsiballZ_command.py 13118 1727204098.69317: Sending initial data 13118 1727204098.69320: Sent initial data (156 bytes) 13118 1727204098.73049: stderr chunk (state=3): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.73053: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.73078: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration <<< 13118 1727204098.73082: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.73269: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204098.73326: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204098.73435: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204098.75185: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13118 1727204098.75211: stderr chunk (state=3): >>>debug1: Using server download size 261120 debug1: Using server upload size 261120 debug1: Server handle limit 1019; using 64 <<< 13118 1727204098.75240: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13118ds5opcmo/tmpxbi4ijoq /root/.ansible/tmp/ansible-tmp-1727204098.6231663-15931-139618413797895/AnsiballZ_command.py <<< 13118 1727204098.75283: stderr chunk (state=3): >>>debug1: Couldn't stat remote file: No such file or directory <<< 13118 1727204098.76606: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204098.76874: stderr chunk (state=3): >>><<< 13118 1727204098.76878: stdout chunk (state=3): >>><<< 13118 1727204098.76880: done transferring module to remote 13118 1727204098.76882: _low_level_execute_command(): starting 13118 1727204098.76885: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727204098.6231663-15931-139618413797895/ /root/.ansible/tmp/ansible-tmp-1727204098.6231663-15931-139618413797895/AnsiballZ_command.py && sleep 0' 13118 1727204098.78227: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204098.78392: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.78411: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.78433: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.78477: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.78488: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204098.78501: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.78517: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204098.78528: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204098.78541: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204098.78551: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.78567: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.78583: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.78594: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.78604: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204098.78616: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.78695: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204098.79388: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204098.79403: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204098.79478: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204098.81307: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204098.81312: stdout chunk (state=3): >>><<< 13118 1727204098.81314: stderr chunk (state=3): >>><<< 13118 1727204098.81413: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204098.81416: _low_level_execute_command(): starting 13118 1727204098.81419: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.9 /root/.ansible/tmp/ansible-tmp-1727204098.6231663-15931-139618413797895/AnsiballZ_command.py && sleep 0' 13118 1727204098.83302: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204098.83327: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.83345: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.83365: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.83466: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.83479: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204098.83492: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.83508: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204098.83518: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204098.83538: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204098.83550: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204098.83566: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204098.83623: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204098.83639: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204098.83653: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204098.83669: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204098.83750: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204098.83887: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204098.83903: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204098.83989: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204099.08148: stdout chunk (state=3): >>> {"changed": true, "stdout": "CHECK DNS AND CONNECTIVITY\n2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org", "stderr": " % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r100 305 100 305 0 0 5446 0 --:--:-- --:--:-- --:--:-- 5446\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r100 291 100 291 0 0 10034 0 --:--:-- --:--:-- --:--:-- 10034", "rc": 0, "cmd": "set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts \"$host\"; then\n echo FAILED to lookup host \"$host\"\n exit 1\n fi\n if ! curl -o /dev/null https://\"$host\"; then\n echo FAILED to contact host \"$host\"\n exit 1\n fi\ndone\n", "start": "2024-09-24 14:54:58.969793", "end": "2024-09-24 14:54:59.080546", "delta": "0:00:00.110753", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts \"$host\"; then\n echo FAILED to lookup host \"$host\"\n exit 1\n fi\n if ! curl -o /dev/null https://\"$host\"; then\n echo FAILED to contact host \"$host\"\n exit 1\n fi\ndone\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13118 1727204099.09489: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. <<< 13118 1727204099.09585: stderr chunk (state=3): >>><<< 13118 1727204099.09590: stdout chunk (state=3): >>><<< 13118 1727204099.09671: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "CHECK DNS AND CONNECTIVITY\n2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org", "stderr": " % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r100 305 100 305 0 0 5446 0 --:--:-- --:--:-- --:--:-- 5446\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r100 291 100 291 0 0 10034 0 --:--:-- --:--:-- --:--:-- 10034", "rc": 0, "cmd": "set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts \"$host\"; then\n echo FAILED to lookup host \"$host\"\n exit 1\n fi\n if ! curl -o /dev/null https://\"$host\"; then\n echo FAILED to contact host \"$host\"\n exit 1\n fi\ndone\n", "start": "2024-09-24 14:54:58.969793", "end": "2024-09-24 14:54:59.080546", "delta": "0:00:00.110753", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts \"$host\"; then\n echo FAILED to lookup host \"$host\"\n exit 1\n fi\n if ! curl -o /dev/null https://\"$host\"; then\n echo FAILED to contact host \"$host\"\n exit 1\n fi\ndone\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.13.78 closed. 13118 1727204099.09800: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts "$host"; then\n echo FAILED to lookup host "$host"\n exit 1\n fi\n if ! curl -o /dev/null https://"$host"; then\n echo FAILED to contact host "$host"\n exit 1\n fi\ndone\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727204098.6231663-15931-139618413797895/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13118 1727204099.09804: _low_level_execute_command(): starting 13118 1727204099.09807: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727204098.6231663-15931-139618413797895/ > /dev/null 2>&1 && sleep 0' 13118 1727204099.11381: stderr chunk (state=2): >>>OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13118 1727204099.11526: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204099.11548: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204099.11576: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204099.11635: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204099.11653: stderr chunk (state=3): >>>debug2: match not found <<< 13118 1727204099.11671: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204099.11690: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13118 1727204099.11703: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.13.78 is address <<< 13118 1727204099.11720: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13118 1727204099.11737: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13118 1727204099.11762: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13118 1727204099.11782: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13118 1727204099.11795: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 <<< 13118 1727204099.11873: stderr chunk (state=3): >>>debug2: match found <<< 13118 1727204099.11895: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13118 1727204099.11991: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master <<< 13118 1727204099.12125: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13118 1727204099.12146: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13118 1727204099.12300: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13118 1727204099.14217: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13118 1727204099.14225: stdout chunk (state=3): >>><<< 13118 1727204099.14227: stderr chunk (state=3): >>><<< 13118 1727204099.14271: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_8.7p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.13.78 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.13.78 originally 10.31.13.78 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13118 1727204099.14275: handler run complete 13118 1727204099.14570: Evaluated conditional (False): False 13118 1727204099.14573: attempt loop complete, returning result 13118 1727204099.14575: _execute() done 13118 1727204099.14578: dumping result to json 13118 1727204099.14580: done dumping result, returning 13118 1727204099.14582: done running TaskExecutor() for managed-node2/TASK: Verify DNS and network connectivity [0affcd87-79f5-56a3-0a64-00000000056e] 13118 1727204099.14584: sending task result for task 0affcd87-79f5-56a3-0a64-00000000056e 13118 1727204099.14749: done sending task result for task 0affcd87-79f5-56a3-0a64-00000000056e 13118 1727204099.14753: WORKER PROCESS EXITING ok: [managed-node2] => { "changed": false, "cmd": "set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts \"$host\"; then\n echo FAILED to lookup host \"$host\"\n exit 1\n fi\n if ! curl -o /dev/null https://\"$host\"; then\n echo FAILED to contact host \"$host\"\n exit 1\n fi\ndone\n", "delta": "0:00:00.110753", "end": "2024-09-24 14:54:59.080546", "rc": 0, "start": "2024-09-24 14:54:58.969793" } STDOUT: CHECK DNS AND CONNECTIVITY 2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.fedoraproject.org 2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.fedoraproject.org 2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.fedoraproject.org 2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.fedoraproject.org 2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.fedoraproject.org 2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.fedoraproject.org 2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org 2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org 2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org 2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org 2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org 2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org STDERR: % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 305 100 305 0 0 5446 0 --:--:-- --:--:-- --:--:-- 5446 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 291 100 291 0 0 10034 0 --:--:-- --:--:-- --:--:-- 10034 13118 1727204099.14829: no more pending results, returning what we have 13118 1727204099.14833: results queue empty 13118 1727204099.14834: checking for any_errors_fatal 13118 1727204099.14844: done checking for any_errors_fatal 13118 1727204099.14845: checking for max_fail_percentage 13118 1727204099.14846: done checking for max_fail_percentage 13118 1727204099.14847: checking to see if all hosts have failed and the running result is not ok 13118 1727204099.14848: done checking to see if all hosts have failed 13118 1727204099.14849: getting the remaining hosts for this loop 13118 1727204099.14851: done getting the remaining hosts for this loop 13118 1727204099.14855: getting the next task for host managed-node2 13118 1727204099.14866: done getting next task for host managed-node2 13118 1727204099.14869: ^ task is: TASK: meta (flush_handlers) 13118 1727204099.14871: ^ state is: HOST STATE: block=3, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204099.14875: getting variables 13118 1727204099.14877: in VariableManager get_vars() 13118 1727204099.14919: Calling all_inventory to load vars for managed-node2 13118 1727204099.14922: Calling groups_inventory to load vars for managed-node2 13118 1727204099.14925: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204099.14936: Calling all_plugins_play to load vars for managed-node2 13118 1727204099.14939: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204099.14942: Calling groups_plugins_play to load vars for managed-node2 13118 1727204099.17576: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204099.20874: done with get_vars() 13118 1727204099.21014: done getting variables 13118 1727204099.21089: in VariableManager get_vars() 13118 1727204099.21106: Calling all_inventory to load vars for managed-node2 13118 1727204099.21108: Calling groups_inventory to load vars for managed-node2 13118 1727204099.21111: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204099.21230: Calling all_plugins_play to load vars for managed-node2 13118 1727204099.21235: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204099.21239: Calling groups_plugins_play to load vars for managed-node2 13118 1727204099.23195: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204099.25225: done with get_vars() 13118 1727204099.25273: done queuing things up, now waiting for results queue to drain 13118 1727204099.25276: results queue empty 13118 1727204099.25277: checking for any_errors_fatal 13118 1727204099.25281: done checking for any_errors_fatal 13118 1727204099.25282: checking for max_fail_percentage 13118 1727204099.25283: done checking for max_fail_percentage 13118 1727204099.25284: checking to see if all hosts have failed and the running result is not ok 13118 1727204099.25285: done checking to see if all hosts have failed 13118 1727204099.25285: getting the remaining hosts for this loop 13118 1727204099.25286: done getting the remaining hosts for this loop 13118 1727204099.25289: getting the next task for host managed-node2 13118 1727204099.25293: done getting next task for host managed-node2 13118 1727204099.25295: ^ task is: TASK: meta (flush_handlers) 13118 1727204099.25412: ^ state is: HOST STATE: block=4, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204099.25422: getting variables 13118 1727204099.25423: in VariableManager get_vars() 13118 1727204099.25443: Calling all_inventory to load vars for managed-node2 13118 1727204099.25446: Calling groups_inventory to load vars for managed-node2 13118 1727204099.25448: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204099.25454: Calling all_plugins_play to load vars for managed-node2 13118 1727204099.25456: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204099.25459: Calling groups_plugins_play to load vars for managed-node2 13118 1727204099.28408: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204099.30290: done with get_vars() 13118 1727204099.30313: done getting variables 13118 1727204099.30487: in VariableManager get_vars() 13118 1727204099.30503: Calling all_inventory to load vars for managed-node2 13118 1727204099.30506: Calling groups_inventory to load vars for managed-node2 13118 1727204099.30508: Calling all_plugins_inventory to load vars for managed-node2 13118 1727204099.30513: Calling all_plugins_play to load vars for managed-node2 13118 1727204099.30516: Calling groups_plugins_inventory to load vars for managed-node2 13118 1727204099.30518: Calling groups_plugins_play to load vars for managed-node2 13118 1727204099.32391: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13118 1727204099.36176: done with get_vars() 13118 1727204099.36441: done queuing things up, now waiting for results queue to drain 13118 1727204099.36444: results queue empty 13118 1727204099.36445: checking for any_errors_fatal 13118 1727204099.36446: done checking for any_errors_fatal 13118 1727204099.36447: checking for max_fail_percentage 13118 1727204099.36448: done checking for max_fail_percentage 13118 1727204099.36448: checking to see if all hosts have failed and the running result is not ok 13118 1727204099.36449: done checking to see if all hosts have failed 13118 1727204099.36449: getting the remaining hosts for this loop 13118 1727204099.36450: done getting the remaining hosts for this loop 13118 1727204099.36453: getting the next task for host managed-node2 13118 1727204099.36456: done getting next task for host managed-node2 13118 1727204099.36457: ^ task is: None 13118 1727204099.36458: ^ state is: HOST STATE: block=5, task=0, rescue=0, always=0, handlers=0, run_state=5, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13118 1727204099.36460: done queuing things up, now waiting for results queue to drain 13118 1727204099.36460: results queue empty 13118 1727204099.36461: checking for any_errors_fatal 13118 1727204099.36462: done checking for any_errors_fatal 13118 1727204099.36462: checking for max_fail_percentage 13118 1727204099.36465: done checking for max_fail_percentage 13118 1727204099.36466: checking to see if all hosts have failed and the running result is not ok 13118 1727204099.36467: done checking to see if all hosts have failed 13118 1727204099.36469: getting the next task for host managed-node2 13118 1727204099.36471: done getting next task for host managed-node2 13118 1727204099.36472: ^ task is: None 13118 1727204099.36473: ^ state is: HOST STATE: block=5, task=0, rescue=0, always=0, handlers=0, run_state=5, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False PLAY RECAP ********************************************************************* managed-node2 : ok=76 changed=3 unreachable=0 failed=0 skipped=60 rescued=0 ignored=0 Tuesday 24 September 2024 14:54:59 -0400 (0:00:00.840) 0:00:36.666 ***** =============================================================================== Install dnsmasq --------------------------------------------------------- 3.63s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:3 Create test interfaces -------------------------------------------------- 1.86s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:35 fedora.linux_system_roles.network : Check which services are running ---- 1.71s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:21 fedora.linux_system_roles.network : Check which services are running ---- 1.64s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:21 Gathering Facts --------------------------------------------------------- 1.45s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tests_bond_nm.yml:6 Install pgrep, sysctl --------------------------------------------------- 1.28s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:26 fedora.linux_system_roles.network : Check which packages are installed --- 1.27s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:26 fedora.linux_system_roles.network : Configure networking connection profiles --- 1.12s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:159 Gathering Facts --------------------------------------------------------- 1.03s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond.yml:3 fedora.linux_system_roles.network : Configure networking connection profiles --- 0.94s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:159 fedora.linux_system_roles.network : Enable and start NetworkManager ----- 0.91s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:122 Verify DNS and network connectivity ------------------------------------- 0.84s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml:24 fedora.linux_system_roles.network : Enable and start NetworkManager ----- 0.82s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:122 fedora.linux_system_roles.network : Check which packages are installed --- 0.80s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:26 Gather the minimum subset of ansible_facts required by the network role test --- 0.79s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:3 Check if system is ostree ----------------------------------------------- 0.74s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:17 fedora.linux_system_roles.network : Re-test connectivity ---------------- 0.66s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:192 fedora.linux_system_roles.network : Re-test connectivity ---------------- 0.52s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:192 Get NM profile info ----------------------------------------------------- 0.52s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:25 Stat profile file ------------------------------------------------------- 0.50s /tmp/collections-G1p/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:9 13118 1727204099.36782: RUNNING CLEANUP