ansible-playbook [core 2.17.7] config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.12/site-packages/ansible ansible collection location = /tmp/collections-BPh executable location = /usr/local/bin/ansible-playbook python version = 3.12.8 (main, Dec 3 2024, 00:00:00) [GCC 14.2.1 20241104 (Red Hat 14.2.1-6)] (/usr/bin/python3.12) jinja version = 3.1.4 libyaml = True No config file found; using defaults running playbook inside collection fedora.linux_system_roles redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug redirecting (type: callback) ansible.builtin.profile_tasks to ansible.posix.profile_tasks Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_basic.yml ****************************************************** 2 plays in /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml PLAY [all] ********************************************************************* TASK [Include vault variables] ************************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:5 Saturday 11 January 2025 11:28:35 -0500 (0:00:00.019) 0:00:00.019 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_test_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n35383939616163653333633431363463313831383037386236646138333162396161356130303461\n3932623930643263313563336163316337643562333936360a363538636631313039343233383732\n38666530383538656639363465313230343533386130303833336434303438333161656262346562\n3362626538613031640a663330613638366132356534363534353239616666653466353961323533\n6565\n" }, "mysql_container_root_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n61333932373230333539663035366431326163363166363036323963623131363530326231303634\n6635326161643165363366323062333334363730376631660a393566366139353861656364656661\n38653463363837336639363032646433666361646535366137303464623261313663643336306465\n6264663730656337310a343962353137386238383064646533366433333437303566656433386233\n34343235326665646661623131643335313236313131353661386338343366316261643634653633\n3832313034366536616531323963333234326461353130303532\n" } }, "ansible_included_var_files": [ "/tmp/podman-9hc/tests/vars/vault-variables.yml" ], "changed": false } PLAY [Ensure that the role runs with default parameters] *********************** TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:9 Saturday 11 January 2025 11:28:35 -0500 (0:00:00.020) 0:00:00.039 ****** [WARNING]: Platform linux on host managed-node2 is using the discovered Python interpreter at /usr/bin/python3.12, but future installation of another Python interpreter could change the meaning of that path. See https://docs.ansible.com/ansible- core/2.17/reference_appendices/interpreter_discovery.html for more information. ok: [managed-node2] TASK [Create tmpdir for testing] *********************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:143 Saturday 11 January 2025 11:28:37 -0500 (0:00:02.451) 0:00:02.491 ****** changed: [managed-node2] => { "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/tmp/lsr_adhv0_3l_podman", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [Change tmpdir permissions] *********************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:150 Saturday 11 January 2025 11:28:37 -0500 (0:00:00.480) 0:00:02.972 ****** changed: [managed-node2] => { "changed": true, "gid": 0, "group": "root", "mode": "0777", "owner": "root", "path": "/tmp/lsr_adhv0_3l_podman", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [Enable podman copr] ****************************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:161 Saturday 11 January 2025 11:28:38 -0500 (0:00:00.447) 0:00:03.420 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [Install podman from updates-testing] ************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:166 Saturday 11 January 2025 11:28:38 -0500 (0:00:00.053) 0:00:03.473 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [Podman version] ********************************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:171 Saturday 11 January 2025 11:28:38 -0500 (0:00:00.052) 0:00:03.525 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [Create user] ************************************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:176 Saturday 11 January 2025 11:28:38 -0500 (0:00:00.051) 0:00:03.576 ****** changed: [managed-node2] => { "changed": true, "comment": "", "create_home": true, "group": 3001, "home": "/home/podman_basic_user", "name": "podman_basic_user", "shell": "/bin/bash", "state": "present", "system": false, "uid": 3001 } TASK [Create tempfile for kube_src] ******************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:181 Saturday 11 January 2025 11:28:39 -0500 (0:00:00.628) 0:00:04.205 ****** changed: [managed-node2 -> localhost] => { "changed": true, "gid": 0, "group": "root", "mode": "0600", "owner": "root", "path": "/tmp/lsr_podman_yvb6_g1x.yml", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Write kube_file_src] ***************************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:189 Saturday 11 January 2025 11:28:39 -0500 (0:00:00.202) 0:00:04.408 ****** changed: [managed-node2 -> localhost] => { "changed": true, "checksum": "347f4489dc195274abbc53ccfc820fe90ed6fb1e", "dest": "/tmp/lsr_podman_yvb6_g1x.yml", "gid": 0, "group": "root", "md5sum": "a45330faac330608948ec69f6ece4970", "mode": "0600", "owner": "root", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 710, "src": "/root/.ansible/tmp/ansible-tmp-1736612919.4404628-7010-169566947857228/.source.yml", "state": "file", "uid": 0 } TASK [Create host directories for data] **************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:196 Saturday 11 January 2025 11:28:40 -0500 (0:00:00.605) 0:00:05.013 ****** changed: [managed-node2] => (item=['httpd1', 'podman_basic_user', 3001]) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": [ "httpd1", "podman_basic_user", 3001 ], "mode": "0755", "owner": "podman_basic_user", "path": "/tmp/lsr_adhv0_3l_podman/httpd1", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 3001 } changed: [managed-node2] => (item=['httpd2', 'root', 0]) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": [ "httpd2", "root", 0 ], "mode": "0755", "owner": "root", "path": "/tmp/lsr_adhv0_3l_podman/httpd2", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } changed: [managed-node2] => (item=['httpd3', 'root', 0]) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": [ "httpd3", "root", 0 ], "mode": "0755", "owner": "root", "path": "/tmp/lsr_adhv0_3l_podman/httpd3", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [Create data files] ******************************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:204 Saturday 11 January 2025 11:28:41 -0500 (0:00:01.080) 0:00:06.094 ****** changed: [managed-node2] => (item=['httpd1', 'podman_basic_user', 3001]) => { "ansible_loop_var": "item", "changed": true, "checksum": "40bd001563085fc35165329ea1ff5c5ecbdbbeef", "dest": "/tmp/lsr_adhv0_3l_podman/httpd1/index.txt", "gid": 0, "group": "root", "item": [ "httpd1", "podman_basic_user", 3001 ], "md5sum": "202cb962ac59075b964b07152d234b70", "mode": "0644", "owner": "podman_basic_user", "secontext": "unconfined_u:object_r:admin_home_t:s0", "size": 3, "src": "/root/.ansible/tmp/ansible-tmp-1736612921.1510544-7094-193496919768865/.source.txt", "state": "file", "uid": 3001 } changed: [managed-node2] => (item=['httpd2', 'root', 0]) => { "ansible_loop_var": "item", "changed": true, "checksum": "40bd001563085fc35165329ea1ff5c5ecbdbbeef", "dest": "/tmp/lsr_adhv0_3l_podman/httpd2/index.txt", "gid": 0, "group": "root", "item": [ "httpd2", "root", 0 ], "md5sum": "202cb962ac59075b964b07152d234b70", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:admin_home_t:s0", "size": 3, "src": "/root/.ansible/tmp/ansible-tmp-1736612921.821283-7094-184250787704089/.source.txt", "state": "file", "uid": 0 } changed: [managed-node2] => (item=['httpd3', 'root', 0]) => { "ansible_loop_var": "item", "changed": true, "checksum": "40bd001563085fc35165329ea1ff5c5ecbdbbeef", "dest": "/tmp/lsr_adhv0_3l_podman/httpd3/index.txt", "gid": 0, "group": "root", "item": [ "httpd3", "root", 0 ], "md5sum": "202cb962ac59075b964b07152d234b70", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:admin_home_t:s0", "size": 3, "src": "/root/.ansible/tmp/ansible-tmp-1736612922.4798434-7094-30966835373969/.source.txt", "state": "file", "uid": 0 } TASK [Run role - do not pull images] ******************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:214 Saturday 11 January 2025 11:28:43 -0500 (0:00:01.999) 0:00:08.093 ****** included: fedora.linux_system_roles.podman for managed-node2 TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Saturday 11 January 2025 11:28:43 -0500 (0:00:00.052) 0:00:08.146 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Saturday 11 January 2025 11:28:43 -0500 (0:00:00.025) 0:00:08.171 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Saturday 11 January 2025 11:28:43 -0500 (0:00:00.034) 0:00:08.206 ****** ok: [managed-node2] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Saturday 11 January 2025 11:28:43 -0500 (0:00:00.353) 0:00:08.560 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Saturday 11 January 2025 11:28:43 -0500 (0:00:00.023) 0:00:08.584 ****** ok: [managed-node2] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Saturday 11 January 2025 11:28:43 -0500 (0:00:00.350) 0:00:08.934 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_is_transactional": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Saturday 11 January 2025 11:28:43 -0500 (0:00:00.023) 0:00:08.957 ****** ok: [managed-node2] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed-node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node2] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } ok: [managed-node2] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Saturday 11 January 2025 11:28:44 -0500 (0:00:00.043) 0:00:09.001 ****** ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Saturday 11 January 2025 11:28:45 -0500 (0:00:00.989) 0:00:09.991 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Saturday 11 January 2025 11:28:45 -0500 (0:00:00.044) 0:00:10.035 ****** changed: [managed-node2] => { "changed": true, "rc": 0, "results": [ "Installed: libgcrypt-1.11.0-5.el10.x86_64", "Installed: shadow-utils-subid-2:4.15.0-5.el10.x86_64", "Installed: libgpg-error-1.50-2.el10.x86_64", "Installed: pcsc-lite-2.2.3-2.el10.x86_64", "Installed: pcsc-lite-ccid-1.6.0-2.el10.x86_64", "Installed: pcsc-lite-libs-2.2.3-2.el10.x86_64", "Installed: npth-1.6-21.el10.x86_64", "Installed: conmon-2:2.1.12-3.el10.x86_64", "Installed: polkit-125-3.el10.x86_64", "Installed: podman-6:5.3.1-3.el10.x86_64", "Installed: polkit-libs-125-3.el10.x86_64", "Installed: libassuan-2.5.6-6.el10.x86_64", "Installed: duktape-2.7.0-10.el10.x86_64", "Installed: polkit-pkla-compat-0.1-30.el10.x86_64", "Installed: gnupg2-smime-2.4.5-2.el10.x86_64", "Installed: container-selinux-3:2.234.2-1.el10.noarch", "Installed: libnet-1.3-7.el10.x86_64", "Installed: libksba-1.6.7-2.el10.x86_64", "Installed: containers-common-5:0.61.0-1.el10.noarch", "Installed: libsecret-0.21.2-7.el10.x86_64", "Installed: protobuf-c-1.5.0-6.el10.x86_64", "Installed: pinentry-1.3.1-1.el10.x86_64", "Installed: containers-common-extra-5:0.61.0-1.el10.noarch", "Installed: gnupg2-2.4.5-2.el10.x86_64", "Installed: gpgme-1.23.2-6.el10.x86_64", "Installed: criu-3.19-6.el10.x86_64", "Installed: criu-libs-3.19-6.el10.x86_64", "Installed: crun-1.19-1.el10.x86_64", "Installed: netavark-2:1.13.1-1.el10.x86_64", "Installed: passt-0^20241121.g238c69f-1.el10.x86_64", "Installed: passt-selinux-0^20241121.g238c69f-1.el10.noarch", "Installed: libusb1-1.0.27-4.el10.x86_64", "Installed: aardvark-dns-2:1.13.1-1.el10.x86_64", "Installed: catatonit-5:0.2.1-1.el10.x86_64" ] } lsrpackages: iptables-nft podman shadow-utils-subid TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Saturday 11 January 2025 11:29:03 -0500 (0:00:18.690) 0:00:28.725 ****** skipping: [managed-node2] => { "false_condition": "__podman_is_transactional | d(false)" } TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Saturday 11 January 2025 11:29:03 -0500 (0:00:00.043) 0:00:28.769 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Saturday 11 January 2025 11:29:03 -0500 (0:00:00.045) 0:00:28.814 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Saturday 11 January 2025 11:29:03 -0500 (0:00:00.041) 0:00:28.855 ****** ok: [managed-node2] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.028347", "end": "2025-01-11 11:29:04.329735", "rc": 0, "start": "2025-01-11 11:29:04.301388" } STDOUT: podman version 5.3.1 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Saturday 11 January 2025 11:29:04 -0500 (0:00:00.530) 0:00:29.386 ****** ok: [managed-node2] => { "ansible_facts": { "podman_version": "5.3.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Saturday 11 January 2025 11:29:04 -0500 (0:00:00.028) 0:00:29.414 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Saturday 11 January 2025 11:29:04 -0500 (0:00:00.028) 0:00:29.443 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "(podman_quadlet_specs | length > 0) or (podman_secrets | length > 0)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Saturday 11 January 2025 11:29:04 -0500 (0:00:00.051) 0:00:29.494 ****** META: end_host conditional evaluated to False, continuing execution for managed-node2 skipping: [managed-node2] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node2" } MSG: end_host conditional evaluated to false, continuing execution for managed-node2 TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Saturday 11 January 2025 11:29:04 -0500 (0:00:00.091) 0:00:29.586 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__has_type_pod or __has_pod_file_ext or __has_pod_file_src_ext or __has_pod_template_src_ext or __has_pod_template_src_ext_j2", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:96 Saturday 11 January 2025 11:29:04 -0500 (0:00:00.054) 0:00:29.641 ****** META: end_host conditional evaluated to False, continuing execution for managed-node2 skipping: [managed-node2] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node2" } MSG: end_host conditional evaluated to false, continuing execution for managed-node2 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:109 Saturday 11 January 2025 11:29:04 -0500 (0:00:00.069) 0:00:29.710 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 11 January 2025 11:29:04 -0500 (0:00:00.053) 0:00:29.764 ****** ok: [managed-node2] => { "ansible_facts": { "getent_passwd": { "root": [ "x", "0", "0", "Super User", "/root", "/bin/bash" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.489) 0:00:30.253 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.030) 0:00:30.283 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.039) 0:00:30.323 ****** ok: [managed-node2] => { "changed": false, "stat": { "atime": 1730678400.0, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1736612940.9742577, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 9160785, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "1643853349", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.375) 0:00:30.698 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.029) 0:00:30.727 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.027) 0:00:30.755 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.028) 0:00:30.783 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.027) 0:00:30.810 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.027) 0:00:30.838 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.027) 0:00:30.865 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.026) 0:00:30.892 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:115 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.026) 0:00:30.918 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:124 Saturday 11 January 2025 11:29:05 -0500 (0:00:00.049) 0:00:30.967 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.049) 0:00:31.017 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.029) 0:00:31.046 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:127 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.029) 0:00:31.076 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.077) 0:00:31.153 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.026) 0:00:31.180 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:130 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.027) 0:00:31.207 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.053) 0:00:31.261 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.028) 0:00:31.289 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:133 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.027) 0:00:31.317 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.054) 0:00:31.372 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.026) 0:00:31.398 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.026) 0:00:31.425 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.028) 0:00:31.453 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:139 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.029) 0:00:31.482 ****** included: fedora.linux_system_roles.firewall for managed-node2 TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.096) 0:00:31.579 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed-node2 TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.049) 0:00:31.629 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Saturday 11 January 2025 11:29:06 -0500 (0:00:00.034) 0:00:31.663 ****** ok: [managed-node2] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Saturday 11 January 2025 11:29:07 -0500 (0:00:00.380) 0:00:32.043 ****** ok: [managed-node2] => { "ansible_facts": { "__firewall_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22 Saturday 11 January 2025 11:29:07 -0500 (0:00:00.035) 0:00:32.079 ****** ok: [managed-node2] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27 Saturday 11 January 2025 11:29:07 -0500 (0:00:00.373) 0:00:32.452 ****** ok: [managed-node2] => { "ansible_facts": { "__firewall_is_transactional": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 Saturday 11 January 2025 11:29:07 -0500 (0:00:00.066) 0:00:32.519 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: firewalld TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43 Saturday 11 January 2025 11:29:08 -0500 (0:00:01.028) 0:00:33.548 ****** skipping: [managed-node2] => { "false_condition": "__firewall_is_transactional | d(false)" } TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48 Saturday 11 January 2025 11:29:08 -0500 (0:00:00.029) 0:00:33.577 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53 Saturday 11 January 2025 11:29:08 -0500 (0:00:00.027) 0:00:33.604 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Saturday 11 January 2025 11:29:08 -0500 (0:00:00.027) 0:00:33.632 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Saturday 11 January 2025 11:29:08 -0500 (0:00:00.025) 0:00:33.658 ****** skipping: [managed-node2] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Saturday 11 January 2025 11:29:08 -0500 (0:00:00.032) 0:00:33.691 ****** ok: [managed-node2] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "system.slice dbus-broker.service dbus.socket sysinit.target basic.target polkit.service", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "network-pre.target shutdown.target", "BindLogSockets": "no", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanLiveMount": "yes", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "ebtables.service ipset.service shutdown.target iptables.service ip6tables.service", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DebugInvocation": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698229248", "EffectiveMemoryMax": "3698229248", "EffectiveTasksMax": "22365", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13978", "LimitNPROCSoft": "13978", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13978", "LimitSIGPENDINGSoft": "13978", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LiveMountResult": "success", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureDurationUSec": "[not set]", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3213991936", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "[not set]", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "[not set]", "MemorySwapMax": "infinity", "MemorySwapPeak": "[not set]", "MemoryZSwapCurrent": "[not set]", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivatePIDs": "no", "PrivateTmp": "no", "PrivateTmpEx": "no", "PrivateUsers": "no", "PrivateUsersEx": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectControlGroupsEx": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "system.slice sysinit.target dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22365", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Saturday 11 January 2025 11:29:09 -0500 (0:00:00.757) 0:00:34.448 ****** changed: [managed-node2] => { "changed": true, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "dbus.socket sysinit.target polkit.service system.slice basic.target dbus-broker.service", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "network-pre.target shutdown.target", "BindLogSockets": "no", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanLiveMount": "yes", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "ipset.service shutdown.target iptables.service ebtables.service ip6tables.service", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DebugInvocation": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698229248", "EffectiveMemoryMax": "3698229248", "EffectiveTasksMax": "22365", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13978", "LimitNPROCSoft": "13978", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13978", "LimitSIGPENDINGSoft": "13978", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LiveMountResult": "success", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureDurationUSec": "[not set]", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3221561344", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "[not set]", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "[not set]", "MemorySwapMax": "infinity", "MemorySwapPeak": "[not set]", "MemoryZSwapCurrent": "[not set]", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivatePIDs": "no", "PrivateTmp": "no", "PrivateTmpEx": "no", "PrivateUsers": "no", "PrivateUsersEx": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectControlGroupsEx": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22365", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Saturday 11 January 2025 11:29:10 -0500 (0:00:01.044) 0:00:35.492 ****** ok: [managed-node2] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3.12", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Saturday 11 January 2025 11:29:10 -0500 (0:00:00.051) 0:00:35.544 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Saturday 11 January 2025 11:29:10 -0500 (0:00:00.031) 0:00:35.576 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Saturday 11 January 2025 11:29:10 -0500 (0:00:00.031) 0:00:35.607 ****** changed: [managed-node2] => (item={'port': '15001-15003/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "15001-15003/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Saturday 11 January 2025 11:29:11 -0500 (0:00:00.718) 0:00:36.325 ****** skipping: [managed-node2] => (item={'port': '15001-15003/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "'detailed' in fw[0]", "item": { "port": "15001-15003/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed-node2] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Saturday 11 January 2025 11:29:11 -0500 (0:00:00.072) 0:00:36.398 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'detailed' in fw[0]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Saturday 11 January 2025 11:29:11 -0500 (0:00:00.046) 0:00:36.444 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Saturday 11 January 2025 11:29:11 -0500 (0:00:00.032) 0:00:36.476 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Saturday 11 January 2025 11:29:11 -0500 (0:00:00.047) 0:00:36.524 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Saturday 11 January 2025 11:29:11 -0500 (0:00:00.046) 0:00:36.570 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Saturday 11 January 2025 11:29:11 -0500 (0:00:00.047) 0:00:36.617 ****** skipping: [managed-node2] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:146 Saturday 11 January 2025 11:29:11 -0500 (0:00:00.071) 0:00:36.689 ****** redirecting (type: modules) ansible.builtin.selinux to ansible.posix.selinux redirecting (type: modules) ansible.builtin.selinux to ansible.posix.selinux redirecting (type: modules) ansible.builtin.seboolean to ansible.posix.seboolean included: fedora.linux_system_roles.selinux for managed-node2 TASK [fedora.linux_system_roles.selinux : Set ansible_facts required by role and install packages] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:2 Saturday 11 January 2025 11:29:12 -0500 (0:00:00.367) 0:00:37.057 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml for managed-node2 TASK [fedora.linux_system_roles.selinux : Ensure ansible_facts used by role] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml:2 Saturday 11 January 2025 11:29:12 -0500 (0:00:00.104) 0:00:37.161 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Ensure SELinux packages] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml:7 Saturday 11 January 2025 11:29:12 -0500 (0:00:00.043) 0:00:37.204 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml for managed-node2 TASK [fedora.linux_system_roles.selinux : Check if system is ostree] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:5 Saturday 11 January 2025 11:29:12 -0500 (0:00:00.066) 0:00:37.270 ****** ok: [managed-node2] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.selinux : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:10 Saturday 11 January 2025 11:29:12 -0500 (0:00:00.420) 0:00:37.691 ****** ok: [managed-node2] => { "ansible_facts": { "__selinux_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.selinux : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:17 Saturday 11 January 2025 11:29:12 -0500 (0:00:00.070) 0:00:37.761 ****** ok: [managed-node2] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.selinux : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:22 Saturday 11 January 2025 11:29:13 -0500 (0:00:00.424) 0:00:38.185 ****** ok: [managed-node2] => { "ansible_facts": { "__selinux_is_transactional": false }, "changed": false } TASK [fedora.linux_system_roles.selinux : Install SELinux python2 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:26 Saturday 11 January 2025 11:29:13 -0500 (0:00:00.040) 0:00:38.226 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_python_version is version('3', '<')", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux python3 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:35 Saturday 11 January 2025 11:29:13 -0500 (0:00:00.061) 0:00:38.287 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: python3-libselinux python3-policycoreutils TASK [fedora.linux_system_roles.selinux : Install SELinux python3 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:46 Saturday 11 January 2025 11:29:14 -0500 (0:00:00.817) 0:00:39.105 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_os_family == \"Suse\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux tool semanage] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:58 Saturday 11 January 2025 11:29:14 -0500 (0:00:00.098) 0:00:39.204 ****** changed: [managed-node2] => { "changed": true, "rc": 0, "results": [ "Installed: policycoreutils-python-utils-3.7-3.el10.noarch" ] } lsrpackages: policycoreutils-python-utils TASK [fedora.linux_system_roles.selinux : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:72 Saturday 11 January 2025 11:29:15 -0500 (0:00:01.185) 0:00:40.389 ****** skipping: [managed-node2] => { "false_condition": "__selinux_is_transactional | d(false)" } TASK [fedora.linux_system_roles.selinux : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:77 Saturday 11 January 2025 11:29:15 -0500 (0:00:00.030) 0:00:40.420 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:82 Saturday 11 January 2025 11:29:15 -0500 (0:00:00.028) 0:00:40.449 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Refresh facts] *********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:89 Saturday 11 January 2025 11:29:15 -0500 (0:00:00.034) 0:00:40.483 ****** ok: [managed-node2] TASK [fedora.linux_system_roles.selinux : Set permanent SELinux state if enabled] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:5 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.945) 0:00:41.429 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_selinux.status == \"enabled\" and (selinux_state or selinux_policy)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set permanent SELinux state if disabled] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:13 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.051) 0:00:41.480 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_selinux.status == \"disabled\" and selinux_state", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set selinux_reboot_required] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:21 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.044) 0:00:41.525 ****** ok: [managed-node2] => { "ansible_facts": { "selinux_reboot_required": false }, "changed": false } TASK [fedora.linux_system_roles.selinux : Fail if reboot is required] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:25 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.104) 0:00:41.629 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_reboot_required", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Warn if SELinux is disabled] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:30 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.045) 0:00:41.675 ****** skipping: [managed-node2] => { "false_condition": "ansible_selinux.status == \"disabled\"" } TASK [fedora.linux_system_roles.selinux : Drop all local modifications] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:35 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.044) 0:00:41.719 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_all_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux boolean local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:43 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.043) 0:00:41.763 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_booleans_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux file context local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:48 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.044) 0:00:41.807 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_fcontexts_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux port local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:53 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.036) 0:00:41.844 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_ports_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux login local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:58 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.030) 0:00:41.875 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_logins_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set SELinux booleans] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:63 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.033) 0:00:41.908 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Set SELinux file contexts] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:74 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.037) 0:00:41.945 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Set an SELinux label on a port] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:87 Saturday 11 January 2025 11:29:16 -0500 (0:00:00.031) 0:00:41.976 ****** changed: [managed-node2] => (item={'ports': '15001-15003', 'setype': 'http_port_t'}) => { "__selinux_item": { "ports": "15001-15003", "setype": "http_port_t" }, "ansible_loop_var": "__selinux_item", "changed": true, "ports": [ "15001-15003" ], "proto": "tcp", "setype": "http_port_t", "state": "present" } TASK [fedora.linux_system_roles.selinux : Set linux user to SELinux user mapping] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:99 Saturday 11 January 2025 11:29:18 -0500 (0:00:01.406) 0:00:43.382 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Get SELinux modules facts] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:112 Saturday 11 January 2025 11:29:18 -0500 (0:00:00.046) 0:00:43.429 ****** ok: [managed-node2] => { "ansible_facts": { "selinux_checksums": true, "selinux_installed_modules": { "abrt": { "100": { "checksum": "sha256:5fdd1acb3c17ce67f51edc9160653adada9aade378218aed1b38acc0b4d4157a", "enabled": 1 } }, "accountsd": { "100": { "checksum": "sha256:50197ed40bbac67de44c12d745bc4a1259c8ae3f3828f5b4cfcd02b187ee1f47", "enabled": 1 } }, "acct": { "100": { "checksum": "sha256:14149edb1593dad6e684a672c57ca35f2ff51aee24ef503171c28671dc1bff13", "enabled": 1 } }, "afs": { "100": { "checksum": "sha256:769aaf25a8699233ceeda87ac858ebc91f094ea98b303c54e2c810760acd9756", "enabled": 1 } }, "afterburn": { "100": { "checksum": "sha256:9b9c00f12dea94989d2f5c592fa8ba4cff49b1ae7c9636fc38baa7957750e80f", "enabled": 1 } }, "aiccu": { "100": { "checksum": "sha256:d7ed1435298abcb42678af33c66ba2bac4c02723601c5018ecef64ca4c4e8f0b", "enabled": 1 } }, "aide": { "100": { "checksum": "sha256:95c6046b0d690df4792fe4418de5f7753279b72c78c5f1dab5397aba8bdf7152", "enabled": 1 } }, "alsa": { "100": { "checksum": "sha256:c0be471cd7182d1126cd6d19b276742a6ace62429425a81d42d8af4c24bcdbb4", "enabled": 1 } }, "amanda": { "100": { "checksum": "sha256:2ea0f86264cad0fa7f1eba0e6238526e92fadc65fd9754f62bc6d59b2da9086c", "enabled": 1 } }, "amtu": { "100": { "checksum": "sha256:1c77cc13007a0024e3121f064aad7dafce5fd57d7f71140f0af6b5af10805a8b", "enabled": 1 } }, "anaconda": { "100": { "checksum": "sha256:79386dbc8f95842f819ddbb01ce5824752d78bb4992783288d79db391a4ec3b5", "enabled": 1 } }, "antivirus": { "100": { "checksum": "sha256:48aafc3e308b4db80797ec75f52d3cc1b8216315d28108a10a03e47a0209adda", "enabled": 1 } }, "apache": { "100": { "checksum": "sha256:30a1e157e56f57a82557524173635e911f796fa8733f98ad190fc854aba893d8", "enabled": 1 } }, "apcupsd": { "100": { "checksum": "sha256:46a2e84d4dd0064f0b368787ed9c494ace4e337d5d80bcc6db77f66dede34375", "enabled": 1 } }, "apm": { "100": { "checksum": "sha256:4f5ff9b6ad61bc6a311bc71768e7ead7fe273989b7cf2fa58b1a13221e8d82f4", "enabled": 1 } }, "application": { "100": { "checksum": "sha256:ee11bb52bc37177920d348c4b920738a1b64413f516c66e18fa5a453a6d81f5e", "enabled": 1 } }, "arpwatch": { "100": { "checksum": "sha256:11a3c765598ae81cda7af561ada4795fb53a9ae9cfcb02af806b08cfc7f00b0e", "enabled": 1 } }, "asterisk": { "100": { "checksum": "sha256:ce887eb790f85810eb5eec0ac96a4e8028d633cf16fc117db4f733e6040c11c2", "enabled": 1 } }, "auditadm": { "100": { "checksum": "sha256:201655fdd63b544aa448c7b46c8b7ec8aa726ef3b55e9ef928ff3502c128973b", "enabled": 1 } }, "authlogin": { "100": { "checksum": "sha256:f1937fa4d112f634f69b9f0104872ce507c56928951a961c26ddf385a12be54e", "enabled": 1 } }, "automount": { "100": { "checksum": "sha256:5d308d3654fc5b0cea77156c27b98bbe2e370b843a920d75e1d616472caee55f", "enabled": 1 } }, "avahi": { "100": { "checksum": "sha256:ab1c62da7d38c1870efff54480b9a43ac2683eb0ebd77925a6fdea0fd47e207c", "enabled": 1 } }, "awstats": { "100": { "checksum": "sha256:f4a83008d582e61321a1d5462807c402d6cd3034d272dab00dd788122f245114", "enabled": 1 } }, "bacula": { "100": { "checksum": "sha256:04e3c063e146bd0f96587e9f49cdb4020e83bf69ba77250d61d2c9fcea53476b", "enabled": 1 } }, "base": { "100": { "checksum": "sha256:106220779e72010bf08e45ebec252a2e47257ca1134c07dbcd68a0d37e790931", "enabled": 1 } }, "bcfg2": { "100": { "checksum": "sha256:860a81a71f68a66e7dc0cee62e02f5da8d2dbb83724bfde9abc24b56b06ec914", "enabled": 1 } }, "bind": { "100": { "checksum": "sha256:ab2767aacd75df866fc6080d7db8dfefbd013a4ad578412ccfa6dece0d0b8ea7", "enabled": 1 } }, "bitlbee": { "100": { "checksum": "sha256:6ae2a654612810131bc874411156b1ef5d784143484e2a867ef413b6ab0f8750", "enabled": 1 } }, "blkmapd": { "100": { "checksum": "sha256:a6ac037c876fcea13d9457717138ec1c17913a0a303469219aa610351225c30c", "enabled": 1 } }, "blueman": { "100": { "checksum": "sha256:10372514dd684076a8cc4c93211d6f6732dcff0fa6aceed02dce868b12aaee82", "enabled": 1 } }, "bluetooth": { "100": { "checksum": "sha256:d8e9f771f666c02cacd7881f83cebd64ca88d79382fddca9e608d75df65099a2", "enabled": 1 } }, "boinc": { "100": { "checksum": "sha256:4ede0a1fb4501325ad8d0ff576f2f20bb0c3e994d2142372449de840f10e8b4b", "enabled": 1 } }, "boltd": { "100": { "checksum": "sha256:63a74e88f61411cfd03aa6e719867d1ae3238b99f3adb2f91cc0973e8b10e174", "enabled": 1 } }, "boothd": { "100": { "checksum": "sha256:4df95e4f92b1da81af245d74fd4481aad048554cad1d31f87985e3c68814ac8e", "enabled": 1 } }, "bootloader": { "100": { "checksum": "sha256:c4b351c2ad386a4c22a477788aeba9ae713cddaae623d733f1957d5d521c717d", "enabled": 1 } }, "bootupd": { "100": { "checksum": "sha256:fd688c68b40c042dd04fe9cb65c7a270353a2f7c677a5e73864b34ee6118f94d", "enabled": 1 } }, "brctl": { "100": { "checksum": "sha256:3a931c1c282cee1b53f6aee3730fef0e498fba3c5005218b4c7a923a7ce226c2", "enabled": 1 } }, "brltty": { "100": { "checksum": "sha256:0471f4ad4a09025a1a23ad398d869d88b90172a869800cf06660c886033c7409", "enabled": 1 } }, "bugzilla": { "100": { "checksum": "sha256:480fe3690a301d2359b747a03b7f65fd881063f6f7d877c55a16e6a51f324df6", "enabled": 1 } }, "cachefilesd": { "100": { "checksum": "sha256:04eafc0b0c7783758bba1cb55893416c204933bf0fdef964ce23c6f17c626c65", "enabled": 1 } }, "calamaris": { "100": { "checksum": "sha256:3b369138db2945167e6b10503d99902411fd5213b91dae978afe344b174240f7", "enabled": 1 } }, "callweaver": { "100": { "checksum": "sha256:0d66c1686efea20f479f6dd3057f5bd5550709d34e3a30745900abf363728bf1", "enabled": 1 } }, "canna": { "100": { "checksum": "sha256:5677c14448ec90dd564b7fe2c19146faa264a6cc80da75dba8f51c1e17af73d2", "enabled": 1 } }, "ccs": { "100": { "checksum": "sha256:ab0abd34c688230d836833a4f46228f3873ccbcfd3fd351d1306c203368ab63b", "enabled": 1 } }, "cdrecord": { "100": { "checksum": "sha256:c258276b231b731865f4237c8477010ea00b71e0e51501d7a206344381f625b1", "enabled": 1 } }, "certmaster": { "100": { "checksum": "sha256:5d46915b4fa32deccdf587b5970cde875d7b43643c09ec42a156193f9592c8ab", "enabled": 1 } }, "certmonger": { "100": { "checksum": "sha256:8746abfbb345c4bfd16b1c27e127a281e72388536aa9c7545ed308a266aceeb5", "enabled": 1 } }, "certwatch": { "100": { "checksum": "sha256:8d8299ad27c884a1657ff8ba52413f24867605e5e31c55a70659b3b0db4309d5", "enabled": 1 } }, "cfengine": { "100": { "checksum": "sha256:c3e0da50f4e8cd55d3a139afdcfa0fc5479b0e4afe78df4a1eab9400a71b64e8", "enabled": 1 } }, "cgroup": { "100": { "checksum": "sha256:e681dc6218ef1387435a6c4e95ed13e45d5edf2089991beb9bc8df41021cd865", "enabled": 1 } }, "chrome": { "100": { "checksum": "sha256:d64d5c77ef6d19be3aecf839fc444fab44a5717fa879c6d1bf6eb7564f556835", "enabled": 1 } }, "chronyd": { "100": { "checksum": "sha256:ea2b23dec628aecd68d6b8bc7d88d662e6bc0e72330f2a87e436f66a38b28ed4", "enabled": 1 } }, "cifsutils": { "100": { "checksum": "sha256:db66a0e17ecf7229eff4ffe84885bbff334085e3e20c282dd7e1eab97545e5e7", "enabled": 1 } }, "cinder": { "100": { "checksum": "sha256:95351f4464622aed05297db68b883d47165ce7db322134048f5a99479c734da3", "enabled": 1 } }, "cipe": { "100": { "checksum": "sha256:25016148cbfa7480cee4b614572bf7a364f6e45a5aca844bb9a87bba217b93a5", "enabled": 1 } }, "clock": { "100": { "checksum": "sha256:5f8b31db8708926fe3de99aadf34b5114184bff20956c7203c6329433bc24dc7", "enabled": 1 } }, "clogd": { "100": { "checksum": "sha256:b6bdd15aeae4574bffcd8a3c862261eb9e21ae5451035cc470ee9eba6cf49cd4", "enabled": 1 } }, "cloudform": { "100": { "checksum": "sha256:8d891dca8241582619e85b4d6f5aa273bd87d3bb7ff04199a38c91f47df12403", "enabled": 1 } }, "cmirrord": { "100": { "checksum": "sha256:e9ccb95f73f750fa464aa4fb5b76e5bad42050e27a128795826272ace5f70760", "enabled": 1 } }, "cobbler": { "100": { "checksum": "sha256:e090b7686819373bcad8a0cdfcf17dd453f7873dfe757742b4bee5c2be7f58d5", "enabled": 1 } }, "collectd": { "100": { "checksum": "sha256:96c655a73312da543bc2aa652b14079cb6fdac677618325d9f1ad83996dbeee1", "enabled": 1 } }, "colord": { "100": { "checksum": "sha256:180a83a6a6059ef9044f9a171ec71a5934a5a19da0f9ce5451c3ba1bbfb75343", "enabled": 1 } }, "comsat": { "100": { "checksum": "sha256:d1e10fcd329ec130a7346374dd6cf339eb38e2c37bc05f371c0381cc7e022d2e", "enabled": 1 } }, "condor": { "100": { "checksum": "sha256:f27781ec40ba6625c3854cf0e6c311a345d5619a3b07c7105635dc60a9369a6c", "enabled": 1 } }, "conman": { "100": { "checksum": "sha256:7984cbaf5572913966bae94db45941e7e15fc195f70c0bd30cbd6b8a24998687", "enabled": 1 } }, "conntrackd": { "100": { "checksum": "sha256:0607c19ac5763806c4420869caf11320ecad25b9ba9e10307105c75e9a26f6be", "enabled": 1 } }, "consolekit": { "100": { "checksum": "sha256:e6e39c814a93fa8973beffb8f891030f7b8094d9987e221d140c1a1fc79a196b", "enabled": 1 } }, "coreos_installer": { "100": { "checksum": "sha256:0ca9390a1bd0c5caa0958c03c1f665a129836f50e3ee11d54fa21800fab9aa08", "enabled": 1 } }, "couchdb": { "100": { "checksum": "sha256:e7ffb1afd19042410ffc4af01ee0fe31ff9f434d23451dd3bd01a7052be716b2", "enabled": 1 } }, "courier": { "100": { "checksum": "sha256:113abb8c428ab21653d9c2f9aef35354470381234ccb493e882ef22f72c2ede4", "enabled": 1 } }, "cpucontrol": { "100": { "checksum": "sha256:36939ecc2515a7980f3be7a2d9518018b3502aa526a446b04552f43a671518dc", "enabled": 1 } }, "cpufreqselector": { "100": { "checksum": "sha256:43b7f9d92ade2e28e26c7f9736d94d91b9a20812118f466b39aa4dc7a02e0959", "enabled": 1 } }, "cpuplug": { "100": { "checksum": "sha256:97a1c9de6df462cab0645b58e16072042c4cab7704c10209950684673a049b8d", "enabled": 1 } }, "cron": { "100": { "checksum": "sha256:0e02e661cc66e751e27d9ece16db633d5ab2036dab470b9dd4a1ef0725bf1534", "enabled": 1 } }, "ctdb": { "100": { "checksum": "sha256:7a7a3998f8a29178005386630556039bba9812f206b5f6782bdcfc0c22f66cbc", "enabled": 1 } }, "cups": { "100": { "checksum": "sha256:90605693f95fba47762dba603d99f0603ccea7238937bd0ed4ca8dc680fd7382", "enabled": 1 } }, "cvs": { "100": { "checksum": "sha256:05274fef587a5ac9def4bce8be2540bc893dd2c7b0052ad9f4e5de0c6da55718", "enabled": 1 } }, "cyphesis": { "100": { "checksum": "sha256:635739376db164b36a935a3bb0dbe8b5644b0fca8d3791b0efde3da36372a4aa", "enabled": 1 } }, "cyrus": { "100": { "checksum": "sha256:26567d47104e4634ab34aa39e0a3ea5f4ed85670b6c66b63e86c47c809080344", "enabled": 1 } }, "daemontools": { "100": { "checksum": "sha256:429c5e1b65232dbfb74741ba1514315288e2704c8a105759c31b85e3742200e1", "enabled": 1 } }, "dbadm": { "100": { "checksum": "sha256:24d6e3dfb505a297328108979f63cbad1b2a0f241c2fb2e84f93dadc086f1273", "enabled": 1 } }, "dbskk": { "100": { "checksum": "sha256:0bf031897570904400ae864c9d12f8880047d1c0368ebb8edc7fd7232f2eecab", "enabled": 1 } }, "dbus": { "100": { "checksum": "sha256:c0b2fa3d02d462ddbf8269485861a3e6e873cd5f05a174a99b121095cbd31f97", "enabled": 1 } }, "dcc": { "100": { "checksum": "sha256:f237213b8df023a69261f26318e3572f28a2c03bccf8d37505dabeca64aa9fe4", "enabled": 1 } }, "ddclient": { "100": { "checksum": "sha256:4159c1afa1a0e1977328780489eb61c30f61823b389f4cd0d79f25a7b27d5857", "enabled": 1 } }, "denyhosts": { "100": { "checksum": "sha256:643ce1a1f7dc2df66e3a1ff57d3dbd9631ea7c9b9b83467d7e4e2f4dc701186b", "enabled": 1 } }, "devicekit": { "100": { "checksum": "sha256:845cf9ac75f705a72df5d2b643b810fe266919369f6840e461f9d3043db97bb9", "enabled": 1 } }, "dhcp": { "100": { "checksum": "sha256:57fe91a9ccded7037429132ca5c28c408881f296eb8e603041b61f82c673c7c6", "enabled": 1 } }, "dictd": { "100": { "checksum": "sha256:246922587bf734f34d26ddb793ab9c649460b3cea63acfca54dc8a989e948d97", "enabled": 1 } }, "dirsrv": { "100": { "checksum": "sha256:9c67bc04c2d69254c0421c864b6f309c4c213bf415759b93a11a21bafa55b7f2", "enabled": 1 } }, "distcc": { "100": { "checksum": "sha256:a8179c2e64684b994ca8802badba7dbe0df42564f400307eeca0274f442cc0c4", "enabled": 1 } }, "dmesg": { "100": { "checksum": "sha256:977de79bba759fc421b791cc5496b0479b34077259eb0b9aab7049cce9e3d537", "enabled": 1 } }, "dmidecode": { "100": { "checksum": "sha256:dbe77a5d3671052e5b1dff63016c24e6e576803c82edbf2a913147a24f893124", "enabled": 1 } }, "dnsmasq": { "100": { "checksum": "sha256:7ad2d3372989319aaef08ac68aff266ae10dd262728284d4531851e893fe9edf", "enabled": 1 } }, "dnssec": { "100": { "checksum": "sha256:70ff660118b3bb9c8e5107d566adf6c2ad0b4af94d6757852c68a00eb19f6e56", "enabled": 1 } }, "dovecot": { "100": { "checksum": "sha256:1057c6ca706e9266c2e147865fb58e634439b7ec3e6c1faba0185401e2d4e985", "enabled": 1 } }, "drbd": { "100": { "checksum": "sha256:225633f600a5c31fcb19a9b721677833e32b88e57ef8cfaa3fa88b25bfc87c87", "enabled": 1 } }, "dspam": { "100": { "checksum": "sha256:503e9b509a0f764faea4a2415311aa1b43bfd8b78bbcc8090f6bc81c200beea8", "enabled": 1 } }, "entropyd": { "100": { "checksum": "sha256:e2aa3a7bc12f543b79842c2504cb1b53b73c55c78a42e24fbead6ab2deb120a4", "enabled": 1 } }, "exim": { "100": { "checksum": "sha256:b097146234c292a604496276a05ae2cbb09bbfdbf0ac9d7948b42afddd953347", "enabled": 1 } }, "fail2ban": { "100": { "checksum": "sha256:c51270c9abffd1d173532bbce08b28480f8e261d099370268b5364c343726140", "enabled": 1 } }, "fcoe": { "100": { "checksum": "sha256:a8a2c86fcc37532b9bf516e070ff53c9e675cb5e9f154966dde6a9b5fc6d7bc5", "enabled": 1 } }, "fdo": { "100": { "checksum": "sha256:7ecdacdeaddd502ea66686e0eaa24fa19fef7202f0092a7174db2aa86eddae8e", "enabled": 1 } }, "fedoratp": { "100": { "checksum": "sha256:85ee3ad2c50bf2b0ba1fb4fe0c57850f9ffd3bb5c11d11a5b5e159ac65783c18", "enabled": 1 } }, "fetchmail": { "100": { "checksum": "sha256:56ef87427ba73e1981192f595dd2c169b322cbed91da8de818e995434e07c757", "enabled": 1 } }, "finger": { "100": { "checksum": "sha256:9aa41b846e607a383c12d0bb4a9695058f8c9b1ed4b26bb59a96855015d9ab35", "enabled": 1 } }, "firewalld": { "100": { "checksum": "sha256:9e85ce564ebec8cae461d093727c4e339a32b10ce3efa569b5c2d53c27baae59", "enabled": 1 } }, "firewallgui": { "100": { "checksum": "sha256:433e16150ab0650ca51df171c1fbe1406cfe15dfa077bade2e48a7d8514bb0ba", "enabled": 1 } }, "firstboot": { "100": { "checksum": "sha256:a16615a4aea8b12fce51f6d7715627da3c3f6c6099db8c3087e29c920d81f861", "enabled": 1 } }, "fprintd": { "100": { "checksum": "sha256:1cb012beb208aebe26b53919c54e8e902210feb7b26d52b05cbe760d273b0645", "enabled": 1 } }, "freeipmi": { "100": { "checksum": "sha256:f62cf10a14cb8de0f158fed8cc9400a8769be1ffb2f3dcdcd0e920de20d68007", "enabled": 1 } }, "freqset": { "100": { "checksum": "sha256:5c96b7146bde0acec28d73afc25b0a35fbfffe2ccd69529717eca6927dc558a7", "enabled": 1 } }, "fstools": { "100": { "checksum": "sha256:0fdc4babffbc8fa18d9445b74b14fc08b77fd6e371c96dc1f5283d6ff7c5f546", "enabled": 1 } }, "ftp": { "100": { "checksum": "sha256:dae3209382de68fdb49a72792108ec3a55c0e50abdfd38c5fae300da4c561c47", "enabled": 1 } }, "fwupd": { "100": { "checksum": "sha256:4f1c84d8ccd64b0af49edc9247ac92f49545fe4dfb24906acc133ebd7ba78bb9", "enabled": 1 } }, "games": { "100": { "checksum": "sha256:7d9b09309f694125d4ba8e1f5683baedfc2c935f0242be94531b43d842509ae9", "enabled": 1 } }, "gdomap": { "100": { "checksum": "sha256:bec0b003cf82cc2fe7e6f196bd82b929e87a827fe8d46bca5b1df2613a409d6b", "enabled": 1 } }, "geoclue": { "100": { "checksum": "sha256:3e8a90189684079f66e782c8e001c435ada771766dd1fa9ae33a977405f6ba99", "enabled": 1 } }, "getty": { "100": { "checksum": "sha256:e5ecf805b71dfe0f039f8eeefa730893c5417fe53b0e0eeba55f99670b4c4584", "enabled": 1 } }, "git": { "100": { "checksum": "sha256:69850d2342f0fcb96a719d39a3698d547c105880e6834f232a4d845a18a046f3", "enabled": 1 } }, "gitosis": { "100": { "checksum": "sha256:24935d3495375c7f6e0a0ddc091cf1f3f276ee155d07760bd7ba0c3ac115f6e8", "enabled": 1 } }, "glance": { "100": { "checksum": "sha256:fe00543de47f940ba5bf1ab84b7dc7f6a680053ef148329f2a9ec827eb47c61c", "enabled": 1 } }, "glusterd": { "100": { "checksum": "sha256:4d2befdcd924820622a0c6653772a9f5ba96aad9a3bd7b4d86daab96b233c9ea", "enabled": 1 } }, "gnome": { "100": { "checksum": "sha256:a5e2adeeda1a5b359025d956cbf20eb3773341207c72cdf7bf5001deeedf4712", "enabled": 1 } }, "gnome_remote_desktop": { "100": { "checksum": "sha256:d926eb26b34a6253b249bc1cea7696dae87887739ef77cd8cb92bdf7f6d1f371", "enabled": 1 } }, "gpg": { "100": { "checksum": "sha256:a9180cca4098d53514cd5f701f4afd40420a3664e8ffa86797ee3a8fadd9a84a", "enabled": 1 } }, "gpm": { "100": { "checksum": "sha256:2b5e83f7513a2470766cbaad46e28604ed8fcd6b640dc0350d90731a807f56ad", "enabled": 1 } }, "gpsd": { "100": { "checksum": "sha256:b64db8adeb9e8602211a0924c1b4b917d0f701df86984e73412344c9b3c2f5d9", "enabled": 1 } }, "gssproxy": { "100": { "checksum": "sha256:a43f376f2b1605cf0d45884b818db045eb37b4f037ae7af36ec5bf993676f4e5", "enabled": 1 } }, "guest": { "100": { "checksum": "sha256:262ccc7c3b237e7ebc31631d83baf2c6cbf8438cce02ce82f008fd6a3ff81a43", "enabled": 1 } }, "hddtemp": { "100": { "checksum": "sha256:fc86a156d59e31b02ce40cbda3c24b0c0baae2ed6f0cabf84cb9a3248ccd236e", "enabled": 1 } }, "hostapd": { "100": { "checksum": "sha256:ba13f04311a741f26ac34cd39688e2f4e75a546358c63c57a2d6a916dfc18315", "enabled": 1 } }, "hostname": { "100": { "checksum": "sha256:7366a48a1774b579953d8b8ae05f9ebe07e5d27984c995afe27509eacf1d79d6", "enabled": 1 } }, "hsqldb": { "100": { "checksum": "sha256:a0c31e8462c7fbbc6579270e5e35923deff56790f565157cb9c32c4eb3d6958b", "enabled": 1 } }, "hwloc": { "100": { "checksum": "sha256:fde041b834e5939945fc0d91e45940745254e128d96625ca5ab4d6e7a4379e27", "enabled": 1 } }, "hypervkvp": { "100": { "checksum": "sha256:520caa8d38f07a36d096f7ee8eb1b0610f49eafedca1721b340f5cdf21e23944", "enabled": 1 } }, "ibacm": { "100": { "checksum": "sha256:4bc0f9a40691f1abb15ce6cd277d7bab4b18720373a197c8310aa63995bcda0f", "enabled": 1 } }, "ica": { "100": { "checksum": "sha256:5daeb045f8b76fed05c1145d131f77dd556de131f777b7e259d8ec57787e7da6", "enabled": 1 } }, "icecast": { "100": { "checksum": "sha256:281ed7601c5b057802da5ffdc24a7e3308e0bb2591a263186f51e35816dffc12", "enabled": 1 } }, "iiosensorproxy": { "100": { "checksum": "sha256:05dd0967ab5bc2422ac57f2ec8a354181bf7038f0d1aabc2775c7c915bbbeaf3", "enabled": 1 } }, "inetd": { "100": { "checksum": "sha256:52f6df31e38b0c20b9fc264115ce56b4d6d8048dce5b6f250a6ca1b0675a0569", "enabled": 1 } }, "init": { "100": { "checksum": "sha256:7141bbe724e0c7f4b3dd7f335ab21267676b126f0443da9cbfa58f1372deb59c", "enabled": 1 } }, "inn": { "100": { "checksum": "sha256:78ae9628e756da216964c88e987dd91e40ec2803ff209763f628a434c8bf3b90", "enabled": 1 } }, "insights_client": { "100": { "checksum": "sha256:c6e547e12ca21f1f68ba74d667fab1bcbfc8d0461fc3a560e80160f32638ddba", "enabled": 1 } }, "iodine": { "100": { "checksum": "sha256:5c8ef0d8e730a3baca51f4c64b13df56149ef304effdc71fe0fb63f90e5a7296", "enabled": 1 } }, "iotop": { "100": { "checksum": "sha256:114de54fd32be9c32e88348ddb168291c0aaaba3a7937501f49e514585228092", "enabled": 1 } }, "ipmievd": { "100": { "checksum": "sha256:5777d2cc5e62942022a8e914847dc7436264585e41e2ce703c49c735e44add8c", "enabled": 1 } }, "ipsec": { "100": { "checksum": "sha256:ffde25d3dd0fa80b933b30cd457ea900389f84d13d4ea629f54c0be59136d4a6", "enabled": 1 } }, "iptables": { "100": { "checksum": "sha256:4c92a7102c209b0edf571a3ca3b35c597bd427d0d258c6829bfd29358448598b", "enabled": 1 } }, "irc": { "100": { "checksum": "sha256:291a822392f8a93d820aa03613bd0c058cfd1118fd6cca93a12175456eb2cf72", "enabled": 1 } }, "irqbalance": { "100": { "checksum": "sha256:2fc17d69e2cba360a0d9043585a3799f86697bf9189fcde62346b41bcf5bd9d9", "enabled": 1 } }, "iscsi": { "100": { "checksum": "sha256:a91d9d18f0502679892047e7a6023107b77412ae4b3d25f75982f766c792be58", "enabled": 1 } }, "isns": { "100": { "checksum": "sha256:5372fd0c432735b6ec28e0a5cafeebd74688912c7703616ca46b77287385c564", "enabled": 1 } }, "jabber": { "100": { "checksum": "sha256:d1916b00ece1c2b546cc04a7b6695f6459724a6bb7a92be78f399f61a613af86", "enabled": 1 } }, "jetty": { "100": { "checksum": "sha256:c78ba53bd379a65017e29895da717fa8e758494e94113b7f324d73121cb35423", "enabled": 1 } }, "jockey": { "100": { "checksum": "sha256:4cee27b59fe57c672587d220f3061417207a02bc0ff36188148f5c5f3101c1ad", "enabled": 1 } }, "journalctl": { "100": { "checksum": "sha256:27aeef17f22b9e891c4c793870f1124ec69ec7dc8f532bfec5b2a0e7319866cb", "enabled": 1 } }, "kafs": { "100": { "checksum": "sha256:602c3c40b28b653e11110e83c7f63122a75b9ad5cb7605299f6951a485284727", "enabled": 1 } }, "kdump": { "100": { "checksum": "sha256:bac33f3ca081f5f4b58fbbdbd3630aef7fb77c19368ec2cef66494cb8ef7fd90", "enabled": 1 } }, "kdumpgui": { "100": { "checksum": "sha256:7e16a86780b3b3d8a6afc5a09335cd6bf6bb39eec53815af9bbc53844655846c", "enabled": 1 } }, "keepalived": { "100": { "checksum": "sha256:5e76b1d8c27eea054956c86458f44ebc46d803035e2b3c0af67fd5feae26b9f2", "enabled": 1 } }, "kerberos": { "100": { "checksum": "sha256:65a81f3d94068c6e7cd3c1092cfbc080f853c7e02490356f7f16d40a3e8ace49", "enabled": 1 } }, "keyboardd": { "100": { "checksum": "sha256:e267709b6fc5f39679c3156b764ac82facb0b228e5b300d891f4a182662a1a5b", "enabled": 1 } }, "keystone": { "100": { "checksum": "sha256:1a7021842ad7350bc5674a455f4232b831ec8153d14635eba8930e8ade0980b7", "enabled": 1 } }, "keyutils": { "100": { "checksum": "sha256:0dc521ba68e3d21396cf5bce20427e6c0c5dfe635c743ccd4374beadd1a96cc5", "enabled": 1 } }, "kismet": { "100": { "checksum": "sha256:4398910d875745fbd61926116eddc0c7f3f6da3cf8b62fb4149d78734c344df3", "enabled": 1 } }, "kpatch": { "100": { "checksum": "sha256:36ae9280a2f7e86befeaf301f1d47087b9214bc97e79562105fcaa1246512467", "enabled": 1 } }, "ksmtuned": { "100": { "checksum": "sha256:66f6fee140f5f767906ea257b32ede5d214b8da0db4031313eb1fddb9a589d21", "enabled": 1 } }, "ktalk": { "100": { "checksum": "sha256:e783786109e8dac25e8be80e5782ec260343cf0db5b03126b85811a101bd9d0e", "enabled": 1 } }, "l2tp": { "100": { "checksum": "sha256:55f189e95389ac2f9205270f3dd53e25761d8e626989bdfb3c05b19c0086dc30", "enabled": 1 } }, "ldap": { "100": { "checksum": "sha256:cb97aa280c3bb13b6c8804e2a2bf971250db72b77ba5ece99702ab6434c3d14a", "enabled": 1 } }, "libraries": { "100": { "checksum": "sha256:bb45e05979c90858a0ed2ed0ecae78cc4e99b750145947a2940eb0493249d161", "enabled": 1 } }, "likewise": { "100": { "checksum": "sha256:f535c1e16dc83389fd5a2fa4716615c687a931e3a9ad7f65dc8607ef305f7994", "enabled": 1 } }, "linuxptp": { "100": { "checksum": "sha256:0d6f39eec4849cc1f5ca0014c29399b78317216ed0e73dfc524be3194ff1bc08", "enabled": 1 } }, "lircd": { "100": { "checksum": "sha256:cdd2ad3d9fa15b37c76e5bf0d781314f3d50897ef6cf377d70706d8a139639f7", "enabled": 1 } }, "livecd": { "100": { "checksum": "sha256:46a683c9c0693f57e0a687926c0ea32a8a176eaa168d1bf86e66463395b314cf", "enabled": 1 } }, "lldpad": { "100": { "checksum": "sha256:9fa26b1dd3ca8e52c93a1320fb66fceb0c8e83730227023d221a6a426b9dddf0", "enabled": 1 } }, "loadkeys": { "100": { "checksum": "sha256:8e876286cd84dd189d40234f691174bc5e0c214056ff0a9f454400cdafdcd6cd", "enabled": 1 } }, "locallogin": { "100": { "checksum": "sha256:f90a7ca564fcff2da3a9effa5c50e8c2a92de3e4c3831f741bc3f8cfc5be5808", "enabled": 1 } }, "lockdev": { "100": { "checksum": "sha256:d181bc52a0d2af9fe776462be417f9f28920a71f01f9894557e9bf7e128ddb5c", "enabled": 1 } }, "logadm": { "100": { "checksum": "sha256:c573637379d69beb37674b0ce9da5f57e60238a88ddad2faa1c3d456cd7df71a", "enabled": 1 } }, "logging": { "100": { "checksum": "sha256:d3494d1da872d6c128a1ab751a7ec84abb1ec73dccbfae200707935617fb8888", "enabled": 1 } }, "logrotate": { "100": { "checksum": "sha256:97dfef99f123dd0474194e78c4e8760df60d46db61f8460161bbcf4b78f7c530", "enabled": 1 } }, "logwatch": { "100": { "checksum": "sha256:c68409b41b76b32647af8f19d1f8f6ebd43430dc820e55d2b094730950a96419", "enabled": 1 } }, "lpd": { "100": { "checksum": "sha256:24594a8f8d0e864712707e8c4114f6d68dd19824045a9f3d10b8f590fc506858", "enabled": 1 } }, "lsm": { "100": { "checksum": "sha256:7df22b87a981a09999b90bb4ecea9713b595814a385fce77eb67e3c20042aa09", "enabled": 1 } }, "lttng-tools": { "100": { "checksum": "sha256:9f30b434e6c08d9a155bbc24ef6d71745d8d252c96d0ded0ccb6d158354c6546", "enabled": 1 } }, "lvm": { "100": { "checksum": "sha256:f0b2c006055d609858e59815fcd002fab6ce4a77761dfd2cc18b56219bf56fc1", "enabled": 1 } }, "mailman": { "100": { "checksum": "sha256:50901ef3e7ccc1828edc3fde37485af69afe28048ba25b7d9e3578b063d559d7", "enabled": 1 } }, "mailscanner": { "100": { "checksum": "sha256:914891424131853f76600695c57ec6283f9cc3c501c06e1427ef99f83431d1cd", "enabled": 1 } }, "man2html": { "100": { "checksum": "sha256:a7af8f4be03248a6d91c5df6c2becd22e3f45a71cecd671d44d2830fcd9f4a3b", "enabled": 1 } }, "mandb": { "100": { "checksum": "sha256:4afad481406029752a7960f97c253b41b8251018b219e16a858ccd029f9bb5f2", "enabled": 1 } }, "mcelog": { "100": { "checksum": "sha256:047b1501df33727c4257a655f1ad55aaed7da1d763782f7da45dfb1b19fa6a02", "enabled": 1 } }, "mediawiki": { "100": { "checksum": "sha256:7ce2990caacbf36b3ea6481def0630fa02e220ea9a4e2dc107251bce1368e569", "enabled": 1 } }, "memcached": { "100": { "checksum": "sha256:577518e3f4a348aaac570906781e3db46d9d53bf52d1a6f83c9f520dab8270e7", "enabled": 1 } }, "milter": { "100": { "checksum": "sha256:aa6812766cbdfad665abf5ad9633ffb945822d31714ffaa2bf4cb061f91e6b6c", "enabled": 1 } }, "minidlna": { "100": { "checksum": "sha256:f13c8b27b597e55d9e6342c756bc4d21de401fcbc4cb01e16c71c01a816c2e2b", "enabled": 1 } }, "minissdpd": { "100": { "checksum": "sha256:18649df6879516ea838eabbd19d6f7aefe2bb7da501c0b471c453d80ea4e02cb", "enabled": 1 } }, "miscfiles": { "100": { "checksum": "sha256:6f5792cde030e56b55f4b00cb53bd6d3724181c80aaaee5806e88c6a5298bb6c", "enabled": 1 } }, "mock": { "100": { "checksum": "sha256:3fdad88af11ebff86688d4db404640a1371af801e4a9b2f14a1b2abc1c76a503", "enabled": 1 } }, "modemmanager": { "100": { "checksum": "sha256:55370db1ae9a5e7ec0b4ec118f45ed61578656c6df3deb7b9e618362ddce4e29", "enabled": 1 } }, "modutils": { "100": { "checksum": "sha256:37b0e80e0d823ecd2606ece25c0b93eaea9f16da35261d79630b3e95c33a5fb4", "enabled": 1 } }, "mojomojo": { "100": { "checksum": "sha256:8ddaca2a54241adf5d42f9205cd073c7062e7692f31ebe15eb9e988e12efc9d4", "enabled": 1 } }, "mon_statd": { "100": { "checksum": "sha256:7e1bf20f24f949d86cf2cf1ecd59aef5aee94c9258d32a8fb4ee7646e4dc67e7", "enabled": 1 } }, "mongodb": { "100": { "checksum": "sha256:051c9eb43df88894edffecf5083ac5065a7051b743068b6f1866d923bca5ab95", "enabled": 1 } }, "motion": { "100": { "checksum": "sha256:154f6949eddb0f008a3b9166f736dde00bbc37df415f07c1263cd63ee609a5fd", "enabled": 1 } }, "mount": { "100": { "checksum": "sha256:acb9c37d19d0aec956cffe6f83d9593cb70bf8ad71ecf956e074b590286f489b", "enabled": 1 } }, "mozilla": { "100": { "checksum": "sha256:0a4982ad6ec48750689dc3c5d7ee946fca86c0c8b3763ee322aae33546109bed", "enabled": 1 } }, "mpd": { "100": { "checksum": "sha256:8807a444e6651ef4e67d0fed3bbc2558dc34b15e7293bfa03487d0a2c54ba905", "enabled": 1 } }, "mplayer": { "100": { "checksum": "sha256:3f999e549842f5a13f5f0d5aa4a752e74216c123f34f1797588b73d504afdc81", "enabled": 1 } }, "mptcpd": { "100": { "checksum": "sha256:f715709d773d8da3d99b4a7de877c7d7a8964d7a1342a61779edcdde2c75702d", "enabled": 1 } }, "mrtg": { "100": { "checksum": "sha256:740151127f1bcdb7699a6317eca571399e7cf58a755afd20dd141aec3035ab53", "enabled": 1 } }, "mta": { "100": { "checksum": "sha256:e04f6d7a528f9eeb7835428fe62d15043b9f4642cded56d682551e18920d04b3", "enabled": 1 } }, "munin": { "100": { "checksum": "sha256:1d2ea7b3403a2c521005ef980913298f01b2c00338366997273344590d0b218e", "enabled": 1 } }, "mysql": { "100": { "checksum": "sha256:07bc412265c70dcb7a3f85ba5265573ef76373364bbbd165ad3abb662e4de705", "enabled": 1 } }, "mythtv": { "100": { "checksum": "sha256:4cab8df79dc5a8e73ece06c418cb3ca58c74f669516b810de46b9f268d78ee8c", "enabled": 1 } }, "nagios": { "100": { "checksum": "sha256:712912fdf95fbf2adc1cb7427524d5536f0cd591bf3e77b63ef069213ae86704", "enabled": 1 } }, "namespace": { "100": { "checksum": "sha256:c032c4722dc3ec48bc9f196ae3de9ef9fd1c04554aa0913d99ade36ac0cf85af", "enabled": 1 } }, "ncftool": { "100": { "checksum": "sha256:e23fc9d15551fc1d7ba386350226d425282fcc090b9dfdcdb7065ebf3385666b", "enabled": 1 } }, "netlabel": { "100": { "checksum": "sha256:d9b7f3b8cb9b0ca564b43893085106dab4a6f6e7bee3143d778fc28218910435", "enabled": 1 } }, "netutils": { "100": { "checksum": "sha256:e828ed9b31086b0b53a5cf66cb2a580aae3e3290156b0d32c07eec5b1bd9cfe5", "enabled": 1 } }, "networkmanager": { "100": { "checksum": "sha256:bc61c1cee0f527dce7154fefb7906e6262e81d84108c7999f8db9cba7e4eb44c", "enabled": 1 } }, "ninfod": { "100": { "checksum": "sha256:4ed636e37871cc6a707bd29f6d990f4ff13d98bac2d8bf37922983d6289b3411", "enabled": 1 } }, "nis": { "100": { "checksum": "sha256:bce01b1750e64ca5f88e7d00d7d4dc88d5ebd430561f9b3cc680484b631c6c8a", "enabled": 1 } }, "nova": { "100": { "checksum": "sha256:16ebf87ef7e152147a0aa89e34161622d387348d4376f2eb91d971ef8ab476a0", "enabled": 1 } }, "nscd": { "100": { "checksum": "sha256:c7e5598679b39499dd00ac032b9a85b6e7e367bc64a0e2827cd799bb7e920a72", "enabled": 1 } }, "nsd": { "100": { "checksum": "sha256:75e019e0b2eda6179a3b39defbd43329dff4627e2c6a06c868ebee8b584332ee", "enabled": 1 } }, "nslcd": { "100": { "checksum": "sha256:69c6eaa2ef51784392d46e962c0f9b9439616dc97ba3105ff54e275a7b1e9db4", "enabled": 1 } }, "ntop": { "100": { "checksum": "sha256:cd4113e904b08b55f2a611ab02edbba284a91097a468b0d6b0d98a5dff397220", "enabled": 1 } }, "ntp": { "100": { "checksum": "sha256:5f188e0c5dabce04e39700cab0a3052cde82676b66e13b054d55f6f78f8d4b30", "enabled": 1 } }, "numad": { "100": { "checksum": "sha256:ed1193b412ae24e79c9a91531c99520b68ded7151bbf0f8cf0353d0dab713dec", "enabled": 1 } }, "nut": { "100": { "checksum": "sha256:54c04020d40b310f4b85f747395b3183859d816a6f181e4a54f5bd6f23b660df", "enabled": 1 } }, "nvme_stas": { "100": { "checksum": "sha256:1098f37f7b13bb6fa26d45bd4c54fd8cb38eddd1b4d02efaf88241f7338022c2", "enabled": 1 } }, "nx": { "100": { "checksum": "sha256:a8b41cc8424e4d9d67310c486221a9fc2701a9508104bdeb076f2ece6c94957c", "enabled": 1 } }, "obex": { "100": { "checksum": "sha256:4810ca21adc14e1f8b8f2a5b91e628038786e71ff84a2b3d701cdcafc6baf16c", "enabled": 1 } }, "oddjob": { "100": { "checksum": "sha256:d09199963de83937d060d3199ad48bf9f6753507c2748ad11d86628469f1e5b5", "enabled": 1 } }, "opafm": { "100": { "checksum": "sha256:020fbfc8bc4b762f26e66c399e2ae1f202c409d7ae1f10ef84f94c9672a1e55f", "enabled": 1 } }, "openct": { "100": { "checksum": "sha256:fd4ba492c3bc1b35ee1d9e4cf2b1fa21ac80ed31c9c76cb4503996f7c6367898", "enabled": 1 } }, "opendnssec": { "100": { "checksum": "sha256:cd5781c1d95be89112ad12c0d1341cc43b60259d4eea9b11442bc19da8ee462c", "enabled": 1 } }, "openfortivpn": { "100": { "checksum": "sha256:9a125b8b0328b3f01aea6aa395ce5117bc3a683e2b45b12e69dc6a95116ce203", "enabled": 1 } }, "openhpid": { "100": { "checksum": "sha256:65ec5e445fa5dd1471b83bccc319e9403a1c5888423442382751b1bf91c5188d", "enabled": 1 } }, "openshift": { "100": { "checksum": "sha256:07bea1192b1c51dd443db5a17d844ba3d7383abc3a3e7e979dcf0143753af074", "enabled": 1 } }, "openshift-origin": { "100": { "checksum": "sha256:88873289a3c1bd0c0c5d636b689bd8e1fe194ca528720c3191fb61b2ca024b50", "enabled": 1 } }, "opensm": { "100": { "checksum": "sha256:76a9f4d16f6b8fe2cb94b30b0c557dd4b067a4cb49502ff812b95290539062f2", "enabled": 1 } }, "openvpn": { "100": { "checksum": "sha256:ec0e41fdd7c04028cedf656a3fd60533adc778663a70734f5f2d569f5e79f53b", "enabled": 1 } }, "openvswitch": { "100": { "checksum": "sha256:1ab8ca3123bd0bbfbe2baac2e44124259401557c590cb8e906b06f28a4b9e32d", "enabled": 1 } }, "openwsman": { "100": { "checksum": "sha256:417b01a65501649535510d72e5fca7174183b5c873c1ddcb6921c052c8be3fc3", "enabled": 1 } }, "oracleasm": { "100": { "checksum": "sha256:6070bb82ad7b693e1ce70bcec6b975c4f5baec8a51e879f27e2c80ff2e076755", "enabled": 1 } }, "osad": { "100": { "checksum": "sha256:7bd10b072510204cc5f32e694b6eb1389fc20c40186fad86b76079e774023f4b", "enabled": 1 } }, "pads": { "100": { "checksum": "sha256:aeae1d20fad2ecc5e2f3ac84829becd50d70a7c41e8183c5b817e0d3f97f9484", "enabled": 1 } }, "passenger": { "100": { "checksum": "sha256:93cb4d1c0096f51325fa10c42ab72db38b5974c42b7d9fa6789482f56ce688c6", "enabled": 1 } }, "pcm": { "100": { "checksum": "sha256:4bb8d2c73a705f3051ce26155740950675bb556b11e2a7f59bb390ebbf7f9a87", "enabled": 1 } }, "pcmcia": { "100": { "checksum": "sha256:722b33d72298e2c8fa2900f1592c92673d098fcb6cd5802449086c918ca104dd", "enabled": 1 } }, "pcp": { "100": { "checksum": "sha256:a34e540529562d554136e3710fe67e8ca6e9ccee02b76dab369a448b58236f34", "enabled": 1 } }, "pcscd": { "100": { "checksum": "sha256:f50088e07d5fd52c3ad752e30852f0d3085cae9f13d2da6fd8c8141f1568ca28", "enabled": 1 } }, "pdns": { "100": { "checksum": "sha256:2d20aff28dd9cac707097c581edf455d754629927a32cc07854eb9baedb23b0a", "enabled": 1 } }, "pegasus": { "100": { "checksum": "sha256:81bedf6564a2a1e2b7ef03a9e5f646184f770202ad403e031f430e494abf8e88", "enabled": 1 } }, "permissivedomains": { "100": { "checksum": "sha256:2453bad4ace526f3cf2c60b358e95a5476692ef25da107b10f52f3af27c056d2", "enabled": 1 } }, "pesign": { "100": { "checksum": "sha256:544140731f1c09b504a6b85ec6ffbdda47c549c9858d6c2264498bfc56fd6127", "enabled": 1 } }, "pingd": { "100": { "checksum": "sha256:a4f1dd19ad8877fb1b48c431790bc350af44d3e2d0d08d59497d2a4425ce2d80", "enabled": 1 } }, "pkcs": { "100": { "checksum": "sha256:c0fe90911ac32abd43e43d9e25dd5a15c85301a1ad0ee579fd3997b9c25ad201", "enabled": 1 } }, "pki": { "100": { "checksum": "sha256:16249004873cf9214f949f73481b5476c0ab03e349542512af9b57aeb637c047", "enabled": 1 } }, "plymouthd": { "100": { "checksum": "sha256:4e33d3ebd3e3cd1fc439ca7bcf3c11086799c7af6a1e12460eb4fc5bab23b0be", "enabled": 1 } }, "podsleuth": { "100": { "checksum": "sha256:7f525e710a117aeaec78a9ab1cbbdd54a150ffe26ed5a1bc29595824fd9e1931", "enabled": 1 } }, "policykit": { "100": { "checksum": "sha256:fe85c9803d13edb28fd5a47803918af2a1e70d66b57e920fefe8e6cfdf778171", "enabled": 1 } }, "polipo": { "100": { "checksum": "sha256:8ab8b8f9438c6ae6c96323b6cf5937528d8223c38629f9bed4a2df00f4b9b686", "enabled": 1 } }, "portmap": { "100": { "checksum": "sha256:59c39477c07ba01841c150a2232e21074575ad3f93782e73b35da8f7acb121af", "enabled": 1 } }, "portreserve": { "100": { "checksum": "sha256:00aa972aa3ea3cf49a0134ebc75df0d6b7eb87ca7b31c61b3107001111abda8a", "enabled": 1 } }, "postfix": { "100": { "checksum": "sha256:68f178fd86c386a5d4df236f983239aa04eb6482304df3e9eb56340b744f5d8c", "enabled": 1 } }, "postgresql": { "100": { "checksum": "sha256:bab7ecb38d7aa40fc1de9d3f1a06e0f83c473e22e2f2391cbdc50e27ef92f7dd", "enabled": 1 } }, "postgrey": { "100": { "checksum": "sha256:8e1915326802644782f3041e65e3b1dfb6f534f2a66f4cb00ca8391f35ef2048", "enabled": 1 } }, "ppp": { "100": { "checksum": "sha256:912c8757c9d3cc46657aa135743c58836db26f4a6abf17807017c86a8512d90f", "enabled": 1 } }, "prelink": { "100": { "checksum": "sha256:0b0ad935c694badc8da9368226c53640cfade72c6828d094924029340ad5521e", "enabled": 1 } }, "prelude": { "100": { "checksum": "sha256:92c5efa0190469791dd3a105bebd81b2891596683001c6cb0d7446fc0d5125a2", "enabled": 1 } }, "privoxy": { "100": { "checksum": "sha256:0c0a1216baa5adcd33f85d92f248893c11cb55b5b068db990b2f94d9d7914932", "enabled": 1 } }, "procmail": { "100": { "checksum": "sha256:ab3962f174f9a9a125e2aad039584a83294b11d3ee56c8b477217a8193afc87d", "enabled": 1 } }, "prosody": { "100": { "checksum": "sha256:3573621c7a830f0d511041287fe8bb073eaa5e200c5b8b2e72096bf469c35f1d", "enabled": 1 } }, "psad": { "100": { "checksum": "sha256:72c77f99e96935aa093669bb095f914484b67e54ae66b4ef8ca780d2758cc6c8", "enabled": 1 } }, "ptchown": { "100": { "checksum": "sha256:318c62c2f63baefa43a61f8ef58bc168596087cd892a63afee72b1cb29cac0c7", "enabled": 1 } }, "pulseaudio": { "100": { "checksum": "sha256:3a7087d4680d1a35893caebf703937f36c1b1799dabc446b02428afc6c163c6f", "enabled": 1 } }, "puppet": { "100": { "checksum": "sha256:366806c68d03db1138cc9a3a08e0fdc4883825c20d1bed9a8e64ce7662bfb032", "enabled": 1 } }, "pwauth": { "100": { "checksum": "sha256:d23c293e39c1a025d375b714dd9319d27545c6c62b58342ef530763b84a24b16", "enabled": 1 } }, "qatlib": { "100": { "checksum": "sha256:cef00aa0bebb89745286c7db05374eb06fe9637c480d295f80f31d21be53672f", "enabled": 1 } }, "qmail": { "100": { "checksum": "sha256:5617289b46ae349fafd315cf92082d64119c6a0b5606bf79aa17fc99e34985b6", "enabled": 1 } }, "qpid": { "100": { "checksum": "sha256:1ee83ce320b057b0569c4bc9f8abdab6233eaf6fcd9c1a78975040bb2d9ddc56", "enabled": 1 } }, "quantum": { "100": { "checksum": "sha256:704e2c7d5fca4cb84e320b9359c6168037ebc87908330c0053c143c8821d3863", "enabled": 1 } }, "quota": { "100": { "checksum": "sha256:58dca8ea36c665ab74eb79cd51d54a7224514660bdfc4b666c2b8031e2877da3", "enabled": 1 } }, "rabbitmq": { "100": { "checksum": "sha256:acc51303259d5e6049265839c5ed13703baa80ae3477ae292103b4665c83dfbb", "enabled": 1 } }, "radius": { "100": { "checksum": "sha256:b53cd6e9f7493004e2b5bdf30e368fbc90b4f971db3fe56e55f31cc839969237", "enabled": 1 } }, "radvd": { "100": { "checksum": "sha256:c909b5d3b8ba285d71358dc7c9fca270e7b1586266c597a60b167ff600d10f7b", "enabled": 1 } }, "raid": { "100": { "checksum": "sha256:071d5ed18ad1c7d334bfba17d392c80a588ad00b1052ab0e21f9c53b0db95e4f", "enabled": 1 } }, "rasdaemon": { "100": { "checksum": "sha256:6d5c3ce954d1cef655ba323f1fef77bda8813815c0d991d020e0c760a4089aa5", "enabled": 1 } }, "rdisc": { "100": { "checksum": "sha256:3b10e49a0fec5da0e8998d77835464b25ced6ee9bd04a24dd861601a9aa02d2d", "enabled": 1 } }, "readahead": { "100": { "checksum": "sha256:e5b3d0b401606c60c1acdfce14c7b6b3f3363a1f58d8d91dd3be36e3df38005f", "enabled": 1 } }, "realmd": { "100": { "checksum": "sha256:b37f4de6bcd25f24eee675c18e2f1842023f0cf9a9a9a7923b64f7eaa875494f", "enabled": 1 } }, "redis": { "100": { "checksum": "sha256:b862f723b5285ad797d5a7546a17b1080690d9218a23b84bb0f7fe618169610f", "enabled": 1 } }, "remotelogin": { "100": { "checksum": "sha256:622d5ae5349b504c7ec9e472a86c5a1e1bc27606868d53d154112a63299c8dc9", "enabled": 1 } }, "restraint": { "400": { "checksum": "sha256:95d0d03fbc1d4147f02a0b3da7cd76efbdd75d1f5812cf6160e922336abbf270", "enabled": 1 } }, "rhcd": { "100": { "checksum": "sha256:fcd8c8415a58d937e726318252127c28e69af2819be01b977b4dda03994834a0", "enabled": 1 } }, "rhcs": { "100": { "checksum": "sha256:1e8298e1e453be52e1a86e391b827f2c71729c066f7d6c63abe7843e28647207", "enabled": 1 } }, "rhev": { "100": { "checksum": "sha256:eaa7e5657902e356ab7700656ff4c185eaf8df675d0a9a3f9ea1a4d8d8fb577d", "enabled": 1 } }, "rhgb": { "100": { "checksum": "sha256:5271469985f35fb7911a72e9e705bc35971294386838f1f11033657874821cea", "enabled": 1 } }, "rhnsd": { "100": { "checksum": "sha256:2a502ce16b7c56df48d0798ed7ca60370d0993ee425a542bf0621b51d4caf2d4", "enabled": 1 } }, "rhsmcertd": { "100": { "checksum": "sha256:80b01c4c3ddf665dd2fd5391b6069e9b868573f7c7a9d4421bc5bfeaadc71a22", "enabled": 1 } }, "rhts": { "400": { "checksum": "sha256:39f15722a115b11064c20b8bc2758e8fe06531a8f923ea00a3e7079a0554e3d6", "enabled": 1 } }, "ricci": { "100": { "checksum": "sha256:dd355c8a0a048baee894e0951365b525bbf5d7cce5907c9919bb0fffb1876661", "enabled": 1 } }, "rkhunter": { "100": { "checksum": "sha256:95ecb10dfe02e0384b0f220c46a7d0c85166a4e2cca0bb55654420b73611a646", "enabled": 1 } }, "rlogin": { "100": { "checksum": "sha256:d8594a73f246468d42569485bb148cbf1b582779074735f6da46e92f763984e1", "enabled": 1 } }, "rngd": { "100": { "checksum": "sha256:50b57a4776d2c6de8b1698bd31ecbfcc98d90d4a9bd2c95672b114c9637ce940", "enabled": 1 } }, "roundup": { "100": { "checksum": "sha256:54c5e85363d326f78c5896bf87c36fca9bc28acee3bd9da4d6d22f090dc9f7c6", "enabled": 1 } }, "rpc": { "100": { "checksum": "sha256:404f69ae03180ddce44c15efc636e5d7ff34b2108cb2c09624075be42a0545b1", "enabled": 1 } }, "rpcbind": { "100": { "checksum": "sha256:f5c703b4107b2616b1adf105b5c592c79f03df1f3a24f31453a242cdec50ea97", "enabled": 1 } }, "rpm": { "100": { "checksum": "sha256:bb52185531861597a0b44bca75497c08ac4c07e10ec63505c333bde800176fb7", "enabled": 1 } }, "rrdcached": { "100": { "checksum": "sha256:2cb11286a0afa94eccbd33de5a32b16d9e178e41ac1fec9dcd801b77dd257051", "enabled": 1 } }, "rshd": { "100": { "checksum": "sha256:0052600f999e54f224d1210c5920b42227cd98ce8db24155b0cfecdf4f6c6474", "enabled": 1 } }, "rshim": { "100": { "checksum": "sha256:7825528bbcdb081ead1d0d2da0d5156e430f93d003f58dcebbb8997c906f0dec", "enabled": 1 } }, "rssh": { "100": { "checksum": "sha256:cde9bfc64b345835e9f8012426cf218bad9972c70e22e0ef6d6a90cc5fa7163c", "enabled": 1 } }, "rsync": { "100": { "checksum": "sha256:1295d1a5434aa1df7fd0f0e2a9b2b045b0ec51bd8619d528349530b322eddf24", "enabled": 1 } }, "rtas": { "100": { "checksum": "sha256:8c5c3a495b42ed14f56e360dd9eaaa0df2426f2c93676880f129bab95e247338", "enabled": 1 } }, "rtkit": { "100": { "checksum": "sha256:7b0c8b0fd23723aed98bc581f2ead99da87893c6ba042ec19ca731e1d802eab6", "enabled": 1 } }, "rwho": { "100": { "checksum": "sha256:fe02b4fb75b8a75ac77b53ae13939199422543f68e71829fd1530aa34c8715cc", "enabled": 1 } }, "samba": { "100": { "checksum": "sha256:b0af5e99697f8e864538813653ea879ea50b6a9269424ca1deaa95ec1a4a7778", "enabled": 1 } }, "sambagui": { "100": { "checksum": "sha256:cde09885f6d657107c517f09f01933816199f0694799fe3df4ac281b411edbb9", "enabled": 1 } }, "sandboxX": { "100": { "checksum": "sha256:ffd79700483dc635b0396df2f9ca75062aff985e7874c2014bca95e80b08bd5d", "enabled": 1 } }, "sanlock": { "100": { "checksum": "sha256:fcd9cea4d61e3597a5e5cb1bd36efc228a21e5b8ee79c68b44e8a93aae834fb6", "enabled": 1 } }, "sap": { "100": { "checksum": "sha256:4cffd87a27c2b48cc59f596a42e860ea52acdfcbff0c2a1e4aeafa3338b29e6e", "enabled": 1 } }, "sasl": { "100": { "checksum": "sha256:ed40cd297b033fd86770692e7d3ab76598684f521f079475d8b44953116291cf", "enabled": 1 } }, "sbd": { "100": { "checksum": "sha256:8c40daa2d8288e7535e9db7724d16a34add080c65fab32bdf5e10cdace95a063", "enabled": 1 } }, "sblim": { "100": { "checksum": "sha256:73b934662023f728bfeb5cac500547e15e389bad3575a3750bf05901d4396531", "enabled": 1 } }, "screen": { "100": { "checksum": "sha256:8dcba8c325fdb2c56f7c7c77362e287469484b19f0b491dd7bc0abc6e033378b", "enabled": 1 } }, "secadm": { "100": { "checksum": "sha256:2d9e07e5ee506dbf6cf976e257210fdb14222c1e437b1f7dceb1548fa399beba", "enabled": 1 } }, "sectoolm": { "100": { "checksum": "sha256:e6c4cabc853877f459381f06d4d47660ce9360a188ab87bb825f1f1eaed39d7d", "enabled": 1 } }, "selinuxutil": { "100": { "checksum": "sha256:38da361f759758e151b7f8612954c91c0fde09615e1c82242c80706231e16355", "enabled": 1 } }, "sendmail": { "100": { "checksum": "sha256:286d18507d80fcfa7cb83ed21321df23c3a29001777890137c1b52c113cd8489", "enabled": 1 } }, "sensord": { "100": { "checksum": "sha256:6681fdfffe9ba81fa3712f665b98feb68f9331f8f282faf19a4ae5e70b02bd63", "enabled": 1 } }, "setrans": { "100": { "checksum": "sha256:69a3bce8a9e35afa4c3b4c7e2f4ba604491313a688d36e3158f96e90343ea370", "enabled": 1 } }, "setroubleshoot": { "100": { "checksum": "sha256:c0d9c0c6cea100697cbc684f2ab74ec66d7c121ca88200752585aeaf4c2b2596", "enabled": 1 } }, "seunshare": { "100": { "checksum": "sha256:219912d16220957b98bbe694f06c3509d2f884ff8418da1eeeda43f4dce81dc9", "enabled": 1 } }, "shorewall": { "100": { "checksum": "sha256:d0dd25c1a5b77a28fcf61f4796c479a026a8e2ec944f5e4d1492cd552aa9ca02", "enabled": 1 } }, "slocate": { "100": { "checksum": "sha256:8d149bca3982ac8c0f7376571d1f0614f05eb2fe7465951eefb17153f642ee40", "enabled": 1 } }, "slpd": { "100": { "checksum": "sha256:6798797c2dda6b8b24aada5a4bbc4f8775faca22435ed1de39193457043f8d8b", "enabled": 1 } }, "smartmon": { "100": { "checksum": "sha256:9de88cb8f60049e56dc5ead706a324189ff276929ccc3bf4ba7d06e96c0d8414", "enabled": 1 } }, "smokeping": { "100": { "checksum": "sha256:29560964ba6e3cb5fc392417ff4aa3e50a5f15c2bc21bdf37088c87868ffeec7", "enabled": 1 } }, "smoltclient": { "100": { "checksum": "sha256:0e47cf6e775a1e2ecb506e444731e086d057b7f2c2aeb6bce7b636372de88d95", "enabled": 1 } }, "snapper": { "100": { "checksum": "sha256:e264d81b3e1173c98c1efc90529d7f9c2c9047b843e29a4fb807d0b1b96b4685", "enabled": 1 } }, "snmp": { "100": { "checksum": "sha256:ec6638b69ec04510a807a5d2c97568a670f80a7999ee711e5e1874136923718e", "enabled": 1 } }, "snort": { "100": { "checksum": "sha256:67c1751c124c249fa15c36f948b60a76db3729f819678cecba569c79abec3af2", "enabled": 1 } }, "sosreport": { "100": { "checksum": "sha256:345531e273b1a0c0d6e02a1898d60759007d5126ec2016eea320832c33ac8211", "enabled": 1 } }, "soundserver": { "100": { "checksum": "sha256:932d162b8365f213b97eb6a7e62d20b8c30bfe1863b6534462da96077d6347fc", "enabled": 1 } }, "spamassassin": { "100": { "checksum": "sha256:fbceb12b459776c22867efa6bf2d07657e058c4a80d6f20e724086ec58e600b0", "enabled": 1 } }, "speech-dispatcher": { "100": { "checksum": "sha256:092593b379ff3e7b3d15b0dfdd2adfde73fb76401d1b6b6e79a4efccbf03f0c9", "enabled": 1 } }, "squid": { "100": { "checksum": "sha256:4374a853cb79bd24d05f65585ad1a74a7b9da4d3454ca233da351e009e75ed79", "enabled": 1 } }, "ssh": { "100": { "checksum": "sha256:493ee98f9cda10ca995eb5600d2940556d70cb3f544501c6bd1968745cc234fa", "enabled": 1 } }, "sslh": { "100": { "checksum": "sha256:5dc03f6d33a6caa7c2fb5ede0f8ded4ecb8eb244443614189cb5909521cab539", "enabled": 1 } }, "sssd": { "100": { "checksum": "sha256:4ee157395107db3b146239f14c9868f6e9d71199520e5d2e473e5e836d0635d1", "enabled": 1 } }, "staff": { "100": { "checksum": "sha256:6a89d52653ce210b1dfb2d2b10e184cacb18976d238fc92b349c40661fdcbf47", "enabled": 1 } }, "stalld": { "100": { "checksum": "sha256:14aa28921885421f92fb51920ca0d8fde122881ac9e28d024c80a7b234b82dd3", "enabled": 1 } }, "stapserver": { "100": { "checksum": "sha256:1610ffc61426c0e791bd7a6a2628935edf3d66fc7818fb42c314044a26ac5d2d", "enabled": 1 } }, "stratisd": { "100": { "checksum": "sha256:5025b28ad7090d4802078c92ba257b46909bbb74fcd6cae0f2a5d65978a5ff8e", "enabled": 1 } }, "stunnel": { "100": { "checksum": "sha256:67b1c960e0cd11cdf1446615cf633700585e69f41c4039cfb183d69527648f1e", "enabled": 1 } }, "su": { "100": { "checksum": "sha256:60850779bd04d036ea792653d4e794f0e865936648c03030515d8125cf412b4b", "enabled": 1 } }, "sudo": { "100": { "checksum": "sha256:0965d3f30b2813a4364e9ce8da0f8916579319ed9507640383c3bf195ae1c00e", "enabled": 1 } }, "svnserve": { "100": { "checksum": "sha256:d589d2e8bddbcea03272bb589cffde776ccb410ac685e5071896fa8e59660c7f", "enabled": 1 } }, "swift": { "100": { "checksum": "sha256:407ddd533381c1f822ec8471e639404c4b926e7e635b3d25d4ba1403a5c517b3", "enabled": 1 } }, "sysadm": { "100": { "checksum": "sha256:1d310d8df59d431579fa3568b7ac14f853223aadd4087fe4c57a1add6606987d", "enabled": 1 } }, "sysadm_secadm": { "100": { "checksum": "sha256:6a0a3b3c4028f304beeacb7d06c62975a2bfb13d67a6a4fe6b95f6d6eeaa74b7", "enabled": 1 } }, "sysnetwork": { "100": { "checksum": "sha256:56b999cfaf7defa2c5cdd5c0c16e7c9c1de940dcc528778ff752fca51f95b38a", "enabled": 1 } }, "sysstat": { "100": { "checksum": "sha256:358b9c77d350c4bc17cf773e99f81d5267f8c41d47402234be37dd9f31655ad0", "enabled": 1 } }, "systemd": { "100": { "checksum": "sha256:124460467ec0f3f7443a67583ef4613154b2b92615e0d779a3ec8060c5a8ce4a", "enabled": 1 } }, "tangd": { "100": { "checksum": "sha256:a2e2a5d807eabe30a09085dcec38d530781c1f02e840a6ed02ccc32bd2738e60", "enabled": 1 } }, "targetd": { "100": { "checksum": "sha256:4a9e67698b4759a406188562c375ecd847f205c38873fb895c3c20abc8bac869", "enabled": 1 } }, "tcpd": { "100": { "checksum": "sha256:b69263d84bf511825be889b667401c30964c7d30d1147d577ae294cd7f05aa01", "enabled": 1 } }, "tcsd": { "100": { "checksum": "sha256:fa40e623b09fb5f61c2c2113a8ed0d17a4a5d105b9f22389ce088a766c0042c0", "enabled": 1 } }, "telepathy": { "100": { "checksum": "sha256:d98021b3bbdd6255c2009e5adafbf41faf3370434396cf2e928e7df693dd50b6", "enabled": 1 } }, "telnet": { "100": { "checksum": "sha256:abf90c09c67553f2b04cab7461138bfd8689752a752a79e36e248f209eae73e6", "enabled": 1 } }, "tftp": { "100": { "checksum": "sha256:f5dcde99c32caa66ea23a9be3b2d6d95550eff7db62b11733f16f1f5eb01bce2", "enabled": 1 } }, "tgtd": { "100": { "checksum": "sha256:6a96f6d7abaa0e1951832347e924c5aff5c66194da3881ec4855e1401a5ce6d2", "enabled": 1 } }, "thin": { "100": { "checksum": "sha256:b0c1055c6a8419035ee9a9a45de2bcdd7183a906bb40d98bb2be80b882a60c51", "enabled": 1 } }, "thumb": { "100": { "checksum": "sha256:928596446716a5d1ef1f8958d150959e59a6258975ebf453c2d5a2747e6d59fd", "enabled": 1 } }, "tlp": { "100": { "checksum": "sha256:540f0a0339feec1e11d145c6ba99f700750451a2fad43190e256e590c3835950", "enabled": 1 } }, "tmpreaper": { "100": { "checksum": "sha256:4370bc58da3c779088151878e7291e03fcd2ed706425b4f2eeeed7097f8adcad", "enabled": 1 } }, "tomcat": { "100": { "checksum": "sha256:98e40c398d94d82eea102055d52f59f77581955b0e463dd95386b4559a4cc852", "enabled": 1 } }, "tor": { "100": { "checksum": "sha256:b309ee2c391d91ace45662d2152870961635ca8968487d0f2570508e160dbd9b", "enabled": 1 } }, "tuned": { "100": { "checksum": "sha256:90bcbb5cad47c032a02390e0c4b22ff79f93fbfd2eb4a385441a87532d6aa632", "enabled": 1 } }, "tvtime": { "100": { "checksum": "sha256:a4438b3063cc1760c663e35a7be70a7dcd099de505ce1562e02090ac6b635787", "enabled": 1 } }, "udev": { "100": { "checksum": "sha256:f308ba2a718171dc871e2586f35a90cb1bbc37decf13e714af72b38cc44505a2", "enabled": 1 } }, "ulogd": { "100": { "checksum": "sha256:05aa3f423e0505dfef9b5e311b0f0a71d93159c7f6440979d3c00a5d9d22ef4c", "enabled": 1 } }, "uml": { "100": { "checksum": "sha256:e0257dd4e250c5aa0baa8151d48942735949bd3018318c11c28a4a5df2aacb90", "enabled": 1 } }, "unconfined": { "100": { "checksum": "sha256:26c4304a314b04f6269b45b4285b449caad98d22de7b3017ef648212c6674da1", "enabled": 1 } }, "unconfineduser": { "100": { "checksum": "sha256:af2778eed6f3cc25c5390e4b1a892995c37a83f5952f51820007b8cd387833b4", "enabled": 1 } }, "unlabelednet": { "100": { "checksum": "sha256:c1de2adfab2cd577a0cc9f2c1a50cd6321324d31b9c3f81be9231f85d4acbbd3", "enabled": 1 } }, "unprivuser": { "100": { "checksum": "sha256:dc1ddbf10a94d55c476ff343c862c24dc649d82190a16a0fd924c527743a667d", "enabled": 1 } }, "updfstab": { "100": { "checksum": "sha256:11eb931e80afefafd0f1c41ea19add0c746170447c76dd629146c43bd9e98067", "enabled": 1 } }, "usbmodules": { "100": { "checksum": "sha256:3a8334855e0cfe67e2c29198baacf61ea0e397d5da30f60fb844044cd105557e", "enabled": 1 } }, "usbmuxd": { "100": { "checksum": "sha256:e883ad3abf2e2792abf4122f99db904139b2f1833b7d99c45b34a7c8244d7c66", "enabled": 1 } }, "userdomain": { "100": { "checksum": "sha256:b6cc74d18c0d17741092aaa8e830f160021ad38bdc38880e3b1488ef30f59a49", "enabled": 1 } }, "userhelper": { "100": { "checksum": "sha256:0a80be6b1635de7593f46421219be5a6ccb9c22973c60b74436ba16884d789fc", "enabled": 1 } }, "usermanage": { "100": { "checksum": "sha256:c56dfa32f163d476dd5761e762ea743189a97636acaf2b4eefcf88812e624877", "enabled": 1 } }, "usernetctl": { "100": { "checksum": "sha256:9c0430aca661392960008075f0a1a934f0c05239976e73c4f0408e9cb5e279ca", "enabled": 1 } }, "uucp": { "100": { "checksum": "sha256:527db920542a250b24e24cb8c6bd86502c3be7e270d11f43d4ecb1a6632f8421", "enabled": 1 } }, "uuidd": { "100": { "checksum": "sha256:b3df65ab2d82cbb05e872f00f46156cfbae937a3a6f850d05f43f4866b772af8", "enabled": 1 } }, "varnishd": { "100": { "checksum": "sha256:9174c3c68bf6b4c2b2b9be7e2a0d985ce6edfdec0dc4c225b427b986449d3e7a", "enabled": 1 } }, "vdagent": { "100": { "checksum": "sha256:52636260af5ae220698d505c6a90a33a71f5878a2cea3a303033c896f000b732", "enabled": 1 } }, "vhostmd": { "100": { "checksum": "sha256:df2d64343b871c840603e82f5fc1f7928c0345b314450e536f9ef66becf131ad", "enabled": 1 } }, "virt": { "100": { "checksum": "sha256:c209f246ed99c66defe424c0f1bd13e970df31d84d2f4e4a2a79ef1c00c3bbe2", "enabled": 1 } }, "virt_supplementary": { "100": { "checksum": "sha256:e8147b1556404b4a2ce521cc4c22e28d4ee42045ea2480043fc8bd2d5a593207", "enabled": 1 } }, "vlock": { "100": { "checksum": "sha256:54a1b28cebc1928622edc4f1f1762a5f2d57783be153052d592f2d10dd9b24a1", "enabled": 1 } }, "vmtools": { "100": { "checksum": "sha256:323b80c9ba1efb9852725d635a5aa442ae2bbd4ec74360cf75853616401dd4f9", "enabled": 1 } }, "vmware": { "100": { "checksum": "sha256:b68b94c1c4662f7005faac011c887619dbb76d69d3d2b63f354d48a144d0d63c", "enabled": 1 } }, "vnstatd": { "100": { "checksum": "sha256:c544da94ee903d411f7ab352b3fe286fa0a1b0fb8490d508565df66d5501b852", "enabled": 1 } }, "vpn": { "100": { "checksum": "sha256:035f214c319dcee2ce4e9751fa9cf1343d81d569409f3a56faf79e506e445504", "enabled": 1 } }, "w3c": { "100": { "checksum": "sha256:d6c6c226dcc92d9a668f42b86dde398d2d875b479915af656734357e20ed0e96", "enabled": 1 } }, "watchdog": { "100": { "checksum": "sha256:a237aeaf94bed19de41ef5289be96ce5bf85dce6f2fc4f9d36d6185b4e59520a", "enabled": 1 } }, "wdmd": { "100": { "checksum": "sha256:1012fdefa5dcf2e540f536de5da79df0ba715ccc2a86a414e2e846675d32495f", "enabled": 1 } }, "webadm": { "100": { "checksum": "sha256:f374fcd23ed5f30ce1159470af1e1254e0352a6ad7cf261287de8ec51f2b2626", "enabled": 1 } }, "webalizer": { "100": { "checksum": "sha256:b66dfdfd23ed304d5f416bb0d8de4565d28cd732253fc2f9e4b4bac59515790f", "enabled": 1 } }, "wine": { "100": { "checksum": "sha256:a4a3be2a0feea4de24ca26c0eed58c7025304d810ee9d4a3e1b7e964a8d28c6d", "enabled": 1 } }, "wireguard": { "100": { "checksum": "sha256:3084b8fd20df45b718cb6779240f4292a0437f536bde7db3616c40b1500ba0f6", "enabled": 1 } }, "wireshark": { "100": { "checksum": "sha256:3ec8e55883a8b73de513f533ed80b1b665e3ed650fdf2b471e5565e1cc3034d9", "enabled": 1 } }, "xen": { "100": { "checksum": "sha256:adabf99dccce9182074a23fb8f05ac537899170fbbfe022f39cb520275339186", "enabled": 1 } }, "xguest": { "100": { "checksum": "sha256:8a7f7d9514fa5257088e451dcf96c6c221a7b6499cae9ac15a688cbe2e84def9", "enabled": 1 } }, "xserver": { "100": { "checksum": "sha256:ca8b72b391320d267580cebf2ad9ef811a781b34647e1c63e525e982911cf956", "enabled": 1 } }, "zabbix": { "100": { "checksum": "sha256:ed9800f41440cf137baa7864acdd8a503b7e051cf023d252b87daeb5357be889", "enabled": 1 } }, "zarafa": { "100": { "checksum": "sha256:c3a9d1c95f377fafb9e94a84bec89645a1053d23b557b143367992eae9993596", "enabled": 1 } }, "zebra": { "100": { "checksum": "sha256:77d57a56a3196f09ea5eb808f8c3b4cef333b241e9007b91779ba78a857e5cfa", "enabled": 1 } }, "zoneminder": { "100": { "checksum": "sha256:675bdf525f863eaf17fe6c25507cf2dc3c4ab05d1e4b0d3477573d34980a5cd2", "enabled": 1 } }, "zosremote": { "100": { "checksum": "sha256:79bb5490f9f16e8caff67fc87a834bf6f11647cfed990c421616eb1adaac5c70", "enabled": 1 } } }, "selinux_priorities": true }, "changed": false } TASK [fedora.linux_system_roles.selinux : Load SELinux modules] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:115 Saturday 11 January 2025 11:29:21 -0500 (0:00:03.315) 0:00:46.745 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_modules is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Restore SELinux labels on filesystem tree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:128 Saturday 11 January 2025 11:29:21 -0500 (0:00:00.059) 0:00:46.804 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Restore SELinux labels on filesystem tree in check mode] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:136 Saturday 11 January 2025 11:29:21 -0500 (0:00:00.057) 0:00:46.862 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:153 Saturday 11 January 2025 11:29:21 -0500 (0:00:00.083) 0:00:46.946 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:157 Saturday 11 January 2025 11:29:22 -0500 (0:00:00.108) 0:00:47.055 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:166 Saturday 11 January 2025 11:29:22 -0500 (0:00:00.057) 0:00:47.112 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:175 Saturday 11 January 2025 11:29:22 -0500 (0:00:00.047) 0:00:47.160 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:182 Saturday 11 January 2025 11:29:22 -0500 (0:00:00.038) 0:00:47.198 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml for managed-node2 => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:14 Saturday 11 January 2025 11:29:22 -0500 (0:00:00.119) 0:00:47.318 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_spec": { "state": "created" }, "__podman_kube_str": "apiVersion: v1\nkind: Pod\nmetadata:\n labels:\n app: test\n io.containers.autoupdate: registry\n name: nopull\nspec:\n containers:\n - name: nopull\n image: quay.io/libpod/testimage:20210610\n" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:21 Saturday 11 January 2025 11:29:22 -0500 (0:00:00.049) 0:00:47.368 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_kube": { "apiVersion": "v1", "kind": "Pod", "metadata": { "labels": { "app": "test", "io.containers.autoupdate": "registry" }, "name": "nopull" }, "spec": { "containers": [ { "image": "quay.io/libpod/testimage:20210610", "name": "nopull" } ] } }, "__podman_kube_file": "", "__podman_pull_image": false, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:33 Saturday 11 January 2025 11:29:22 -0500 (0:00:00.052) 0:00:47.420 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_name": "nopull", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:38 Saturday 11 January 2025 11:29:22 -0500 (0:00:00.039) 0:00:47.460 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 11 January 2025 11:29:22 -0500 (0:00:00.140) 0:00:47.600 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 11 January 2025 11:29:22 -0500 (0:00:00.102) 0:00:47.703 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 11 January 2025 11:29:22 -0500 (0:00:00.053) 0:00:47.756 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Saturday 11 January 2025 11:29:22 -0500 (0:00:00.056) 0:00:47.813 ****** ok: [managed-node2] => { "changed": false, "stat": { "atime": 1736612945.6352425, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1736612940.9742577, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 9160785, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "1643853349", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.391) 0:00:48.205 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.038) 0:00:48.243 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.047) 0:00:48.291 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.063) 0:00:48.355 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.061) 0:00:48.417 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.050) 0:00:48.467 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.046) 0:00:48.514 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.045) 0:00:48.559 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if no kube spec is given] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:43 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.055) 0:00:48.614 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_kube", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:52 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.046) 0:00:48.660 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_activate_systemd_unit": false, "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:60 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.097) 0:00:48.757 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_path": "/etc/containers/ansible-kubernetes.d" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:64 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.046) 0:00:48.804 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_file": "/etc/containers/ansible-kubernetes.d/nopull.yml" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:68 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.069) 0:00:48.873 ****** ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Get service name using systemd-escape] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:75 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.049) 0:00:48.922 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_activate_systemd_unit | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Cleanup containers and services] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:83 Saturday 11 January 2025 11:29:23 -0500 (0:00:00.049) 0:00:48.972 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update containers and services] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:87 Saturday 11 January 2025 11:29:24 -0500 (0:00:00.041) 0:00:49.014 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:2 Saturday 11 January 2025 11:29:24 -0500 (0:00:00.083) 0:00:49.098 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 11 January 2025 11:29:24 -0500 (0:00:00.062) 0:00:49.161 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 11 January 2025 11:29:24 -0500 (0:00:00.035) 0:00:49.196 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 11 January 2025 11:29:24 -0500 (0:00:00.039) 0:00:49.236 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the host mount volumes] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:7 Saturday 11 January 2025 11:29:24 -0500 (0:00:00.089) 0:00:49.326 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'volumes' in __podman_kube['spec']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:18 Saturday 11 January 2025 11:29:24 -0500 (0:00:00.041) 0:00:49.368 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_volumes | d([]) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:29 Saturday 11 January 2025 11:29:24 -0500 (0:00:00.038) 0:00:49.406 ****** skipping: [managed-node2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Check the kubernetes yaml file] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:53 Saturday 11 January 2025 11:29:24 -0500 (0:00:00.052) 0:00:49.459 ****** ok: [managed-node2] => { "changed": false, "failed_when_result": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Ensure the kubernetes directory is present] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:61 Saturday 11 January 2025 11:29:24 -0500 (0:00:00.386) 0:00:49.845 ****** changed: [managed-node2] => { "changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/ansible-kubernetes.d", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure kubernetes yaml files are present] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:70 Saturday 11 January 2025 11:29:25 -0500 (0:00:00.390) 0:00:50.236 ****** changed: [managed-node2] => { "changed": true, "checksum": "d5dc917e3cae36de03aa971a17ac473f86fdf934", "dest": "/etc/containers/ansible-kubernetes.d/nopull.yml", "gid": 0, "group": "root", "md5sum": "1eceaf0da0bbf69a778deb11f0449417", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 217, "src": "/root/.ansible/tmp/ansible-tmp-1736612965.2957008-8350-171865569281770/.source.yml", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Update containers/pods] *************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:80 Saturday 11 January 2025 11:29:25 -0500 (0:00:00.696) 0:00:50.932 ****** [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed-node2] => { "actions": [ "/usr/bin/podman play kube --start=false /etc/containers/ansible-kubernetes.d/nopull.yml" ], "changed": true } STDOUT: Pod: 8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844 Container: 010c205b58090946e480f7d230e44e257dd9db66fbcfe0636fb2c68d69f1bba4 STDERR: Trying to pull quay.io/libpod/testimage:20210610... Getting image source signatures Copying blob sha256:9afcdfe780b4ea44cc52d22e3f93ccf212388a90370773571ce034a62e14174e Copying config sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f Writing manifest to image destination TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:88 Saturday 11 January 2025 11:29:29 -0500 (0:00:03.068) 0:00:54.000 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_activate_systemd_unit | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Enable service] *********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:100 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.042) 0:00:54.042 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_activate_systemd_unit | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:113 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.046) 0:00:54.089 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_activate_systemd_unit | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:127 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.043) 0:00:54.132 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_activate_systemd_unit | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:189 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.042) 0:00:54.174 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:196 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.073) 0:00:54.248 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle credential files - absent] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:202 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.033) 0:00:54.282 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:211 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.032) 0:00:54.314 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [Verify image not pulled] ************************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:235 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.053) 0:00:54.368 ****** ok: [managed-node2] => { "changed": false } MSG: All assertions passed TASK [Run role - verify continue if pull fails] ******************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:239 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.044) 0:00:54.412 ****** included: fedora.linux_system_roles.podman for managed-node2 TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.103) 0:00:54.515 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.064) 0:00:54.579 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.060) 0:00:54.640 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.066) 0:00:54.706 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.120) 0:00:54.827 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.068) 0:00:54.895 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Saturday 11 January 2025 11:29:29 -0500 (0:00:00.080) 0:00:54.976 ****** ok: [managed-node2] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed-node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node2] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } ok: [managed-node2] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Saturday 11 January 2025 11:29:30 -0500 (0:00:00.126) 0:00:55.103 ****** ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Saturday 11 January 2025 11:29:31 -0500 (0:00:00.907) 0:00:56.011 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Saturday 11 January 2025 11:29:31 -0500 (0:00:00.052) 0:00:56.063 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages))", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Saturday 11 January 2025 11:29:31 -0500 (0:00:00.052) 0:00:56.115 ****** skipping: [managed-node2] => { "false_condition": "__podman_is_transactional | d(false)" } TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Saturday 11 January 2025 11:29:31 -0500 (0:00:00.037) 0:00:56.153 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Saturday 11 January 2025 11:29:31 -0500 (0:00:00.037) 0:00:56.191 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Saturday 11 January 2025 11:29:31 -0500 (0:00:00.037) 0:00:56.228 ****** ok: [managed-node2] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.024745", "end": "2025-01-11 11:29:31.583308", "rc": 0, "start": "2025-01-11 11:29:31.558563" } STDOUT: podman version 5.3.1 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Saturday 11 January 2025 11:29:31 -0500 (0:00:00.460) 0:00:56.689 ****** ok: [managed-node2] => { "ansible_facts": { "podman_version": "5.3.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Saturday 11 January 2025 11:29:31 -0500 (0:00:00.054) 0:00:56.743 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Saturday 11 January 2025 11:29:31 -0500 (0:00:00.059) 0:00:56.803 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "(podman_quadlet_specs | length > 0) or (podman_secrets | length > 0)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Saturday 11 January 2025 11:29:31 -0500 (0:00:00.043) 0:00:56.847 ****** META: end_host conditional evaluated to False, continuing execution for managed-node2 skipping: [managed-node2] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node2" } MSG: end_host conditional evaluated to false, continuing execution for managed-node2 TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Saturday 11 January 2025 11:29:31 -0500 (0:00:00.057) 0:00:56.904 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__has_type_pod or __has_pod_file_ext or __has_pod_file_src_ext or __has_pod_template_src_ext or __has_pod_template_src_ext_j2", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:96 Saturday 11 January 2025 11:29:31 -0500 (0:00:00.052) 0:00:56.957 ****** META: end_host conditional evaluated to False, continuing execution for managed-node2 skipping: [managed-node2] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node2" } MSG: end_host conditional evaluated to false, continuing execution for managed-node2 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:109 Saturday 11 January 2025 11:29:32 -0500 (0:00:00.054) 0:00:57.012 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 11 January 2025 11:29:32 -0500 (0:00:00.098) 0:00:57.111 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 11 January 2025 11:29:32 -0500 (0:00:00.058) 0:00:57.170 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 11 January 2025 11:29:32 -0500 (0:00:00.135) 0:00:57.306 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Saturday 11 January 2025 11:29:32 -0500 (0:00:00.098) 0:00:57.404 ****** ok: [managed-node2] => { "changed": false, "stat": { "atime": 1736612945.6352425, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1736612940.9742577, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 9160785, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "1643853349", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Saturday 11 January 2025 11:29:32 -0500 (0:00:00.419) 0:00:57.823 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Saturday 11 January 2025 11:29:32 -0500 (0:00:00.042) 0:00:57.866 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Saturday 11 January 2025 11:29:32 -0500 (0:00:00.041) 0:00:57.908 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Saturday 11 January 2025 11:29:32 -0500 (0:00:00.042) 0:00:57.950 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.055) 0:00:58.005 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.061) 0:00:58.067 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.079) 0:00:58.147 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.052) 0:00:58.200 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:115 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.045) 0:00:58.246 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:124 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.046) 0:00:58.292 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.116) 0:00:58.408 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.045) 0:00:58.453 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:127 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.045) 0:00:58.499 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.073) 0:00:58.573 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.039) 0:00:58.612 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:130 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.037) 0:00:58.650 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.076) 0:00:58.727 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.065) 0:00:58.792 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:133 Saturday 11 January 2025 11:29:33 -0500 (0:00:00.065) 0:00:58.857 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Saturday 11 January 2025 11:29:34 -0500 (0:00:00.144) 0:00:59.001 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Saturday 11 January 2025 11:29:34 -0500 (0:00:00.042) 0:00:59.044 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Saturday 11 January 2025 11:29:34 -0500 (0:00:00.037) 0:00:59.081 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Saturday 11 January 2025 11:29:34 -0500 (0:00:00.036) 0:00:59.118 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:139 Saturday 11 January 2025 11:29:34 -0500 (0:00:00.038) 0:00:59.156 ****** included: fedora.linux_system_roles.firewall for managed-node2 TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Saturday 11 January 2025 11:29:34 -0500 (0:00:00.171) 0:00:59.328 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed-node2 TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Saturday 11 January 2025 11:29:34 -0500 (0:00:00.087) 0:00:59.415 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Saturday 11 January 2025 11:29:34 -0500 (0:00:00.054) 0:00:59.469 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Saturday 11 January 2025 11:29:34 -0500 (0:00:00.090) 0:00:59.560 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22 Saturday 11 January 2025 11:29:34 -0500 (0:00:00.041) 0:00:59.602 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27 Saturday 11 January 2025 11:29:34 -0500 (0:00:00.045) 0:00:59.647 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 Saturday 11 January 2025 11:29:34 -0500 (0:00:00.046) 0:00:59.694 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: firewalld TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43 Saturday 11 January 2025 11:29:35 -0500 (0:00:00.752) 0:01:00.446 ****** skipping: [managed-node2] => { "false_condition": "__firewall_is_transactional | d(false)" } TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48 Saturday 11 January 2025 11:29:35 -0500 (0:00:00.065) 0:01:00.512 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53 Saturday 11 January 2025 11:29:35 -0500 (0:00:00.048) 0:01:00.561 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Saturday 11 January 2025 11:29:35 -0500 (0:00:00.055) 0:01:00.617 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Saturday 11 January 2025 11:29:35 -0500 (0:00:00.043) 0:01:00.660 ****** skipping: [managed-node2] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Saturday 11 January 2025 11:29:35 -0500 (0:00:00.047) 0:01:00.708 ****** ok: [managed-node2] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sat 2025-01-11 11:29:10 EST", "ActiveEnterTimestampMonotonic": "271605433", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus-broker.service dbus.socket sysinit.target polkit.service system.slice basic.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sat 2025-01-11 11:29:10 EST", "AssertTimestampMonotonic": "271333792", "Before": "multi-user.target shutdown.target network-pre.target", "BindLogSockets": "no", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "398429000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanLiveMount": "yes", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2025-01-11 11:29:10 EST", "ConditionTimestampMonotonic": "271333789", "ConfigurationDirectoryMode": "0755", "Conflicts": "iptables.service ipset.service ip6tables.service ebtables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4516", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DebugInvocation": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698229248", "EffectiveMemoryMax": "3698229248", "EffectiveTasksMax": "22365", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainHandoffTimestampMonotonic": "271363582", "ExecMainPID": "10862", "ExecMainStartTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainStartTimestampMonotonic": "271336631", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2025-01-11 11:29:10 EST", "InactiveExitTimestampMonotonic": "271337679", "InvocationID": "f2f268445be048bc90d6f558b31f85a8", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13978", "LimitNPROCSoft": "13978", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13978", "LimitSIGPENDINGSoft": "13978", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LiveMountResult": "success", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "10862", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureDurationUSec": "[not set]", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3242127360", "MemoryCurrent": "32956416", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "35106816", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "0", "MemorySwapMax": "infinity", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivatePIDs": "no", "PrivateTmp": "no", "PrivateTmpEx": "no", "PrivateUsers": "no", "PrivateUsersEx": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectControlGroupsEx": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket sysinit.target system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestamp": "Sat 2025-01-11 11:29:10 EST", "StateChangeTimestampMonotonic": "271605433", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22365", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Saturday 11 January 2025 11:29:36 -0500 (0:00:00.610) 0:01:01.319 ****** ok: [managed-node2] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sat 2025-01-11 11:29:10 EST", "ActiveEnterTimestampMonotonic": "271605433", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus-broker.service dbus.socket sysinit.target polkit.service system.slice basic.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sat 2025-01-11 11:29:10 EST", "AssertTimestampMonotonic": "271333792", "Before": "multi-user.target shutdown.target network-pre.target", "BindLogSockets": "no", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "398429000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanLiveMount": "yes", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2025-01-11 11:29:10 EST", "ConditionTimestampMonotonic": "271333789", "ConfigurationDirectoryMode": "0755", "Conflicts": "iptables.service ipset.service ip6tables.service ebtables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4516", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DebugInvocation": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698229248", "EffectiveMemoryMax": "3698229248", "EffectiveTasksMax": "22365", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainHandoffTimestampMonotonic": "271363582", "ExecMainPID": "10862", "ExecMainStartTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainStartTimestampMonotonic": "271336631", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2025-01-11 11:29:10 EST", "InactiveExitTimestampMonotonic": "271337679", "InvocationID": "f2f268445be048bc90d6f558b31f85a8", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13978", "LimitNPROCSoft": "13978", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13978", "LimitSIGPENDINGSoft": "13978", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LiveMountResult": "success", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "10862", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureDurationUSec": "[not set]", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3225538560", "MemoryCurrent": "32956416", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "35106816", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "0", "MemorySwapMax": "infinity", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivatePIDs": "no", "PrivateTmp": "no", "PrivateTmpEx": "no", "PrivateUsers": "no", "PrivateUsersEx": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectControlGroupsEx": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket sysinit.target system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestamp": "Sat 2025-01-11 11:29:10 EST", "StateChangeTimestampMonotonic": "271605433", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22365", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Saturday 11 January 2025 11:29:36 -0500 (0:00:00.554) 0:01:01.874 ****** ok: [managed-node2] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3.12", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Saturday 11 January 2025 11:29:36 -0500 (0:00:00.097) 0:01:01.972 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Saturday 11 January 2025 11:29:37 -0500 (0:00:00.043) 0:01:02.015 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Saturday 11 January 2025 11:29:37 -0500 (0:00:00.063) 0:01:02.079 ****** ok: [managed-node2] => (item={'port': '15001-15003/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "15001-15003/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Saturday 11 January 2025 11:29:37 -0500 (0:00:00.560) 0:01:02.640 ****** skipping: [managed-node2] => (item={'port': '15001-15003/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "'detailed' in fw[0]", "item": { "port": "15001-15003/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed-node2] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Saturday 11 January 2025 11:29:37 -0500 (0:00:00.097) 0:01:02.737 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'detailed' in fw[0]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Saturday 11 January 2025 11:29:37 -0500 (0:00:00.088) 0:01:02.825 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Saturday 11 January 2025 11:29:37 -0500 (0:00:00.070) 0:01:02.896 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Saturday 11 January 2025 11:29:37 -0500 (0:00:00.070) 0:01:02.966 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Saturday 11 January 2025 11:29:38 -0500 (0:00:00.066) 0:01:03.033 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Saturday 11 January 2025 11:29:38 -0500 (0:00:00.061) 0:01:03.094 ****** skipping: [managed-node2] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:146 Saturday 11 January 2025 11:29:38 -0500 (0:00:00.070) 0:01:03.165 ****** redirecting (type: modules) ansible.builtin.selinux to ansible.posix.selinux redirecting (type: modules) ansible.builtin.selinux to ansible.posix.selinux redirecting (type: modules) ansible.builtin.seboolean to ansible.posix.seboolean included: fedora.linux_system_roles.selinux for managed-node2 TASK [fedora.linux_system_roles.selinux : Set ansible_facts required by role and install packages] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:2 Saturday 11 January 2025 11:29:38 -0500 (0:00:00.193) 0:01:03.358 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml for managed-node2 TASK [fedora.linux_system_roles.selinux : Ensure ansible_facts used by role] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml:2 Saturday 11 January 2025 11:29:38 -0500 (0:00:00.101) 0:01:03.460 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Ensure SELinux packages] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml:7 Saturday 11 January 2025 11:29:38 -0500 (0:00:00.053) 0:01:03.514 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml for managed-node2 TASK [fedora.linux_system_roles.selinux : Check if system is ostree] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:5 Saturday 11 January 2025 11:29:38 -0500 (0:00:00.083) 0:01:03.597 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:10 Saturday 11 January 2025 11:29:38 -0500 (0:00:00.043) 0:01:03.641 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:17 Saturday 11 January 2025 11:29:38 -0500 (0:00:00.041) 0:01:03.683 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:22 Saturday 11 January 2025 11:29:38 -0500 (0:00:00.041) 0:01:03.724 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux python2 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:26 Saturday 11 January 2025 11:29:38 -0500 (0:00:00.042) 0:01:03.767 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_python_version is version('3', '<')", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux python3 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:35 Saturday 11 January 2025 11:29:38 -0500 (0:00:00.044) 0:01:03.811 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: python3-libselinux python3-policycoreutils TASK [fedora.linux_system_roles.selinux : Install SELinux python3 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:46 Saturday 11 January 2025 11:29:39 -0500 (0:00:00.821) 0:01:04.633 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_os_family == \"Suse\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux tool semanage] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:58 Saturday 11 January 2025 11:29:39 -0500 (0:00:00.071) 0:01:04.704 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: policycoreutils-python-utils TASK [fedora.linux_system_roles.selinux : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:72 Saturday 11 January 2025 11:29:40 -0500 (0:00:00.780) 0:01:05.485 ****** skipping: [managed-node2] => { "false_condition": "__selinux_is_transactional | d(false)" } TASK [fedora.linux_system_roles.selinux : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:77 Saturday 11 January 2025 11:29:40 -0500 (0:00:00.049) 0:01:05.535 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:82 Saturday 11 January 2025 11:29:40 -0500 (0:00:00.041) 0:01:05.576 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Refresh facts] *********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:89 Saturday 11 January 2025 11:29:40 -0500 (0:00:00.038) 0:01:05.615 ****** ok: [managed-node2] TASK [fedora.linux_system_roles.selinux : Set permanent SELinux state if enabled] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:5 Saturday 11 January 2025 11:29:41 -0500 (0:00:00.877) 0:01:06.492 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_selinux.status == \"enabled\" and (selinux_state or selinux_policy)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set permanent SELinux state if disabled] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:13 Saturday 11 January 2025 11:29:41 -0500 (0:00:00.090) 0:01:06.583 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_selinux.status == \"disabled\" and selinux_state", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set selinux_reboot_required] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:21 Saturday 11 January 2025 11:29:41 -0500 (0:00:00.065) 0:01:06.649 ****** ok: [managed-node2] => { "ansible_facts": { "selinux_reboot_required": false }, "changed": false } TASK [fedora.linux_system_roles.selinux : Fail if reboot is required] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:25 Saturday 11 January 2025 11:29:41 -0500 (0:00:00.078) 0:01:06.728 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_reboot_required", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Warn if SELinux is disabled] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:30 Saturday 11 January 2025 11:29:41 -0500 (0:00:00.061) 0:01:06.790 ****** skipping: [managed-node2] => { "false_condition": "ansible_selinux.status == \"disabled\"" } TASK [fedora.linux_system_roles.selinux : Drop all local modifications] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:35 Saturday 11 January 2025 11:29:41 -0500 (0:00:00.062) 0:01:06.852 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_all_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux boolean local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:43 Saturday 11 January 2025 11:29:41 -0500 (0:00:00.129) 0:01:06.982 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_booleans_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux file context local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:48 Saturday 11 January 2025 11:29:42 -0500 (0:00:00.058) 0:01:07.041 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_fcontexts_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux port local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:53 Saturday 11 January 2025 11:29:42 -0500 (0:00:00.043) 0:01:07.084 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_ports_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux login local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:58 Saturday 11 January 2025 11:29:42 -0500 (0:00:00.047) 0:01:07.132 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_logins_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set SELinux booleans] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:63 Saturday 11 January 2025 11:29:42 -0500 (0:00:00.040) 0:01:07.172 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Set SELinux file contexts] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:74 Saturday 11 January 2025 11:29:42 -0500 (0:00:00.036) 0:01:07.208 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Set an SELinux label on a port] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:87 Saturday 11 January 2025 11:29:42 -0500 (0:00:00.035) 0:01:07.244 ****** ok: [managed-node2] => (item={'ports': '15001-15003', 'setype': 'http_port_t'}) => { "__selinux_item": { "ports": "15001-15003", "setype": "http_port_t" }, "ansible_loop_var": "__selinux_item", "changed": false, "ports": [ "15001-15003" ], "proto": "tcp", "setype": "http_port_t", "state": "present" } TASK [fedora.linux_system_roles.selinux : Set linux user to SELinux user mapping] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:99 Saturday 11 January 2025 11:29:42 -0500 (0:00:00.609) 0:01:07.854 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Get SELinux modules facts] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:112 Saturday 11 January 2025 11:29:42 -0500 (0:00:00.062) 0:01:07.916 ****** ok: [managed-node2] => { "ansible_facts": { "selinux_checksums": true, "selinux_installed_modules": { "abrt": { "100": { "checksum": "sha256:5fdd1acb3c17ce67f51edc9160653adada9aade378218aed1b38acc0b4d4157a", "enabled": 1 } }, "accountsd": { "100": { "checksum": "sha256:50197ed40bbac67de44c12d745bc4a1259c8ae3f3828f5b4cfcd02b187ee1f47", "enabled": 1 } }, "acct": { "100": { "checksum": "sha256:14149edb1593dad6e684a672c57ca35f2ff51aee24ef503171c28671dc1bff13", "enabled": 1 } }, "afs": { "100": { "checksum": "sha256:769aaf25a8699233ceeda87ac858ebc91f094ea98b303c54e2c810760acd9756", "enabled": 1 } }, "afterburn": { "100": { "checksum": "sha256:9b9c00f12dea94989d2f5c592fa8ba4cff49b1ae7c9636fc38baa7957750e80f", "enabled": 1 } }, "aiccu": { "100": { "checksum": "sha256:d7ed1435298abcb42678af33c66ba2bac4c02723601c5018ecef64ca4c4e8f0b", "enabled": 1 } }, "aide": { "100": { "checksum": "sha256:95c6046b0d690df4792fe4418de5f7753279b72c78c5f1dab5397aba8bdf7152", "enabled": 1 } }, "alsa": { "100": { "checksum": "sha256:c0be471cd7182d1126cd6d19b276742a6ace62429425a81d42d8af4c24bcdbb4", "enabled": 1 } }, "amanda": { "100": { "checksum": "sha256:2ea0f86264cad0fa7f1eba0e6238526e92fadc65fd9754f62bc6d59b2da9086c", "enabled": 1 } }, "amtu": { "100": { "checksum": "sha256:1c77cc13007a0024e3121f064aad7dafce5fd57d7f71140f0af6b5af10805a8b", "enabled": 1 } }, "anaconda": { "100": { "checksum": "sha256:79386dbc8f95842f819ddbb01ce5824752d78bb4992783288d79db391a4ec3b5", "enabled": 1 } }, "antivirus": { "100": { "checksum": "sha256:48aafc3e308b4db80797ec75f52d3cc1b8216315d28108a10a03e47a0209adda", "enabled": 1 } }, "apache": { "100": { "checksum": "sha256:30a1e157e56f57a82557524173635e911f796fa8733f98ad190fc854aba893d8", "enabled": 1 } }, "apcupsd": { "100": { "checksum": "sha256:46a2e84d4dd0064f0b368787ed9c494ace4e337d5d80bcc6db77f66dede34375", "enabled": 1 } }, "apm": { "100": { "checksum": "sha256:4f5ff9b6ad61bc6a311bc71768e7ead7fe273989b7cf2fa58b1a13221e8d82f4", "enabled": 1 } }, "application": { "100": { "checksum": "sha256:ee11bb52bc37177920d348c4b920738a1b64413f516c66e18fa5a453a6d81f5e", "enabled": 1 } }, "arpwatch": { "100": { "checksum": "sha256:11a3c765598ae81cda7af561ada4795fb53a9ae9cfcb02af806b08cfc7f00b0e", "enabled": 1 } }, "asterisk": { "100": { "checksum": "sha256:ce887eb790f85810eb5eec0ac96a4e8028d633cf16fc117db4f733e6040c11c2", "enabled": 1 } }, "auditadm": { "100": { "checksum": "sha256:201655fdd63b544aa448c7b46c8b7ec8aa726ef3b55e9ef928ff3502c128973b", "enabled": 1 } }, "authlogin": { "100": { "checksum": "sha256:f1937fa4d112f634f69b9f0104872ce507c56928951a961c26ddf385a12be54e", "enabled": 1 } }, "automount": { "100": { "checksum": "sha256:5d308d3654fc5b0cea77156c27b98bbe2e370b843a920d75e1d616472caee55f", "enabled": 1 } }, "avahi": { "100": { "checksum": "sha256:ab1c62da7d38c1870efff54480b9a43ac2683eb0ebd77925a6fdea0fd47e207c", "enabled": 1 } }, "awstats": { "100": { "checksum": "sha256:f4a83008d582e61321a1d5462807c402d6cd3034d272dab00dd788122f245114", "enabled": 1 } }, "bacula": { "100": { "checksum": "sha256:04e3c063e146bd0f96587e9f49cdb4020e83bf69ba77250d61d2c9fcea53476b", "enabled": 1 } }, "base": { "100": { "checksum": "sha256:106220779e72010bf08e45ebec252a2e47257ca1134c07dbcd68a0d37e790931", "enabled": 1 } }, "bcfg2": { "100": { "checksum": "sha256:860a81a71f68a66e7dc0cee62e02f5da8d2dbb83724bfde9abc24b56b06ec914", "enabled": 1 } }, "bind": { "100": { "checksum": "sha256:ab2767aacd75df866fc6080d7db8dfefbd013a4ad578412ccfa6dece0d0b8ea7", "enabled": 1 } }, "bitlbee": { "100": { "checksum": "sha256:6ae2a654612810131bc874411156b1ef5d784143484e2a867ef413b6ab0f8750", "enabled": 1 } }, "blkmapd": { "100": { "checksum": "sha256:a6ac037c876fcea13d9457717138ec1c17913a0a303469219aa610351225c30c", "enabled": 1 } }, "blueman": { "100": { "checksum": "sha256:10372514dd684076a8cc4c93211d6f6732dcff0fa6aceed02dce868b12aaee82", "enabled": 1 } }, "bluetooth": { "100": { "checksum": "sha256:d8e9f771f666c02cacd7881f83cebd64ca88d79382fddca9e608d75df65099a2", "enabled": 1 } }, "boinc": { "100": { "checksum": "sha256:4ede0a1fb4501325ad8d0ff576f2f20bb0c3e994d2142372449de840f10e8b4b", "enabled": 1 } }, "boltd": { "100": { "checksum": "sha256:63a74e88f61411cfd03aa6e719867d1ae3238b99f3adb2f91cc0973e8b10e174", "enabled": 1 } }, "boothd": { "100": { "checksum": "sha256:4df95e4f92b1da81af245d74fd4481aad048554cad1d31f87985e3c68814ac8e", "enabled": 1 } }, "bootloader": { "100": { "checksum": "sha256:c4b351c2ad386a4c22a477788aeba9ae713cddaae623d733f1957d5d521c717d", "enabled": 1 } }, "bootupd": { "100": { "checksum": "sha256:fd688c68b40c042dd04fe9cb65c7a270353a2f7c677a5e73864b34ee6118f94d", "enabled": 1 } }, "brctl": { "100": { "checksum": "sha256:3a931c1c282cee1b53f6aee3730fef0e498fba3c5005218b4c7a923a7ce226c2", "enabled": 1 } }, "brltty": { "100": { "checksum": "sha256:0471f4ad4a09025a1a23ad398d869d88b90172a869800cf06660c886033c7409", "enabled": 1 } }, "bugzilla": { "100": { "checksum": "sha256:480fe3690a301d2359b747a03b7f65fd881063f6f7d877c55a16e6a51f324df6", "enabled": 1 } }, "cachefilesd": { "100": { "checksum": "sha256:04eafc0b0c7783758bba1cb55893416c204933bf0fdef964ce23c6f17c626c65", "enabled": 1 } }, "calamaris": { "100": { "checksum": "sha256:3b369138db2945167e6b10503d99902411fd5213b91dae978afe344b174240f7", "enabled": 1 } }, "callweaver": { "100": { "checksum": "sha256:0d66c1686efea20f479f6dd3057f5bd5550709d34e3a30745900abf363728bf1", "enabled": 1 } }, "canna": { "100": { "checksum": "sha256:5677c14448ec90dd564b7fe2c19146faa264a6cc80da75dba8f51c1e17af73d2", "enabled": 1 } }, "ccs": { "100": { "checksum": "sha256:ab0abd34c688230d836833a4f46228f3873ccbcfd3fd351d1306c203368ab63b", "enabled": 1 } }, "cdrecord": { "100": { "checksum": "sha256:c258276b231b731865f4237c8477010ea00b71e0e51501d7a206344381f625b1", "enabled": 1 } }, "certmaster": { "100": { "checksum": "sha256:5d46915b4fa32deccdf587b5970cde875d7b43643c09ec42a156193f9592c8ab", "enabled": 1 } }, "certmonger": { "100": { "checksum": "sha256:8746abfbb345c4bfd16b1c27e127a281e72388536aa9c7545ed308a266aceeb5", "enabled": 1 } }, "certwatch": { "100": { "checksum": "sha256:8d8299ad27c884a1657ff8ba52413f24867605e5e31c55a70659b3b0db4309d5", "enabled": 1 } }, "cfengine": { "100": { "checksum": "sha256:c3e0da50f4e8cd55d3a139afdcfa0fc5479b0e4afe78df4a1eab9400a71b64e8", "enabled": 1 } }, "cgroup": { "100": { "checksum": "sha256:e681dc6218ef1387435a6c4e95ed13e45d5edf2089991beb9bc8df41021cd865", "enabled": 1 } }, "chrome": { "100": { "checksum": "sha256:d64d5c77ef6d19be3aecf839fc444fab44a5717fa879c6d1bf6eb7564f556835", "enabled": 1 } }, "chronyd": { "100": { "checksum": "sha256:ea2b23dec628aecd68d6b8bc7d88d662e6bc0e72330f2a87e436f66a38b28ed4", "enabled": 1 } }, "cifsutils": { "100": { "checksum": "sha256:db66a0e17ecf7229eff4ffe84885bbff334085e3e20c282dd7e1eab97545e5e7", "enabled": 1 } }, "cinder": { "100": { "checksum": "sha256:95351f4464622aed05297db68b883d47165ce7db322134048f5a99479c734da3", "enabled": 1 } }, "cipe": { "100": { "checksum": "sha256:25016148cbfa7480cee4b614572bf7a364f6e45a5aca844bb9a87bba217b93a5", "enabled": 1 } }, "clock": { "100": { "checksum": "sha256:5f8b31db8708926fe3de99aadf34b5114184bff20956c7203c6329433bc24dc7", "enabled": 1 } }, "clogd": { "100": { "checksum": "sha256:b6bdd15aeae4574bffcd8a3c862261eb9e21ae5451035cc470ee9eba6cf49cd4", "enabled": 1 } }, "cloudform": { "100": { "checksum": "sha256:8d891dca8241582619e85b4d6f5aa273bd87d3bb7ff04199a38c91f47df12403", "enabled": 1 } }, "cmirrord": { "100": { "checksum": "sha256:e9ccb95f73f750fa464aa4fb5b76e5bad42050e27a128795826272ace5f70760", "enabled": 1 } }, "cobbler": { "100": { "checksum": "sha256:e090b7686819373bcad8a0cdfcf17dd453f7873dfe757742b4bee5c2be7f58d5", "enabled": 1 } }, "collectd": { "100": { "checksum": "sha256:96c655a73312da543bc2aa652b14079cb6fdac677618325d9f1ad83996dbeee1", "enabled": 1 } }, "colord": { "100": { "checksum": "sha256:180a83a6a6059ef9044f9a171ec71a5934a5a19da0f9ce5451c3ba1bbfb75343", "enabled": 1 } }, "comsat": { "100": { "checksum": "sha256:d1e10fcd329ec130a7346374dd6cf339eb38e2c37bc05f371c0381cc7e022d2e", "enabled": 1 } }, "condor": { "100": { "checksum": "sha256:f27781ec40ba6625c3854cf0e6c311a345d5619a3b07c7105635dc60a9369a6c", "enabled": 1 } }, "conman": { "100": { "checksum": "sha256:7984cbaf5572913966bae94db45941e7e15fc195f70c0bd30cbd6b8a24998687", "enabled": 1 } }, "conntrackd": { "100": { "checksum": "sha256:0607c19ac5763806c4420869caf11320ecad25b9ba9e10307105c75e9a26f6be", "enabled": 1 } }, "consolekit": { "100": { "checksum": "sha256:e6e39c814a93fa8973beffb8f891030f7b8094d9987e221d140c1a1fc79a196b", "enabled": 1 } }, "coreos_installer": { "100": { "checksum": "sha256:0ca9390a1bd0c5caa0958c03c1f665a129836f50e3ee11d54fa21800fab9aa08", "enabled": 1 } }, "couchdb": { "100": { "checksum": "sha256:e7ffb1afd19042410ffc4af01ee0fe31ff9f434d23451dd3bd01a7052be716b2", "enabled": 1 } }, "courier": { "100": { "checksum": "sha256:113abb8c428ab21653d9c2f9aef35354470381234ccb493e882ef22f72c2ede4", "enabled": 1 } }, "cpucontrol": { "100": { "checksum": "sha256:36939ecc2515a7980f3be7a2d9518018b3502aa526a446b04552f43a671518dc", "enabled": 1 } }, "cpufreqselector": { "100": { "checksum": "sha256:43b7f9d92ade2e28e26c7f9736d94d91b9a20812118f466b39aa4dc7a02e0959", "enabled": 1 } }, "cpuplug": { "100": { "checksum": "sha256:97a1c9de6df462cab0645b58e16072042c4cab7704c10209950684673a049b8d", "enabled": 1 } }, "cron": { "100": { "checksum": "sha256:0e02e661cc66e751e27d9ece16db633d5ab2036dab470b9dd4a1ef0725bf1534", "enabled": 1 } }, "ctdb": { "100": { "checksum": "sha256:7a7a3998f8a29178005386630556039bba9812f206b5f6782bdcfc0c22f66cbc", "enabled": 1 } }, "cups": { "100": { "checksum": "sha256:90605693f95fba47762dba603d99f0603ccea7238937bd0ed4ca8dc680fd7382", "enabled": 1 } }, "cvs": { "100": { "checksum": "sha256:05274fef587a5ac9def4bce8be2540bc893dd2c7b0052ad9f4e5de0c6da55718", "enabled": 1 } }, "cyphesis": { "100": { "checksum": "sha256:635739376db164b36a935a3bb0dbe8b5644b0fca8d3791b0efde3da36372a4aa", "enabled": 1 } }, "cyrus": { "100": { "checksum": "sha256:26567d47104e4634ab34aa39e0a3ea5f4ed85670b6c66b63e86c47c809080344", "enabled": 1 } }, "daemontools": { "100": { "checksum": "sha256:429c5e1b65232dbfb74741ba1514315288e2704c8a105759c31b85e3742200e1", "enabled": 1 } }, "dbadm": { "100": { "checksum": "sha256:24d6e3dfb505a297328108979f63cbad1b2a0f241c2fb2e84f93dadc086f1273", "enabled": 1 } }, "dbskk": { "100": { "checksum": "sha256:0bf031897570904400ae864c9d12f8880047d1c0368ebb8edc7fd7232f2eecab", "enabled": 1 } }, "dbus": { "100": { "checksum": "sha256:c0b2fa3d02d462ddbf8269485861a3e6e873cd5f05a174a99b121095cbd31f97", "enabled": 1 } }, "dcc": { "100": { "checksum": "sha256:f237213b8df023a69261f26318e3572f28a2c03bccf8d37505dabeca64aa9fe4", "enabled": 1 } }, "ddclient": { "100": { "checksum": "sha256:4159c1afa1a0e1977328780489eb61c30f61823b389f4cd0d79f25a7b27d5857", "enabled": 1 } }, "denyhosts": { "100": { "checksum": "sha256:643ce1a1f7dc2df66e3a1ff57d3dbd9631ea7c9b9b83467d7e4e2f4dc701186b", "enabled": 1 } }, "devicekit": { "100": { "checksum": "sha256:845cf9ac75f705a72df5d2b643b810fe266919369f6840e461f9d3043db97bb9", "enabled": 1 } }, "dhcp": { "100": { "checksum": "sha256:57fe91a9ccded7037429132ca5c28c408881f296eb8e603041b61f82c673c7c6", "enabled": 1 } }, "dictd": { "100": { "checksum": "sha256:246922587bf734f34d26ddb793ab9c649460b3cea63acfca54dc8a989e948d97", "enabled": 1 } }, "dirsrv": { "100": { "checksum": "sha256:9c67bc04c2d69254c0421c864b6f309c4c213bf415759b93a11a21bafa55b7f2", "enabled": 1 } }, "distcc": { "100": { "checksum": "sha256:a8179c2e64684b994ca8802badba7dbe0df42564f400307eeca0274f442cc0c4", "enabled": 1 } }, "dmesg": { "100": { "checksum": "sha256:977de79bba759fc421b791cc5496b0479b34077259eb0b9aab7049cce9e3d537", "enabled": 1 } }, "dmidecode": { "100": { "checksum": "sha256:dbe77a5d3671052e5b1dff63016c24e6e576803c82edbf2a913147a24f893124", "enabled": 1 } }, "dnsmasq": { "100": { "checksum": "sha256:7ad2d3372989319aaef08ac68aff266ae10dd262728284d4531851e893fe9edf", "enabled": 1 } }, "dnssec": { "100": { "checksum": "sha256:70ff660118b3bb9c8e5107d566adf6c2ad0b4af94d6757852c68a00eb19f6e56", "enabled": 1 } }, "dovecot": { "100": { "checksum": "sha256:1057c6ca706e9266c2e147865fb58e634439b7ec3e6c1faba0185401e2d4e985", "enabled": 1 } }, "drbd": { "100": { "checksum": "sha256:225633f600a5c31fcb19a9b721677833e32b88e57ef8cfaa3fa88b25bfc87c87", "enabled": 1 } }, "dspam": { "100": { "checksum": "sha256:503e9b509a0f764faea4a2415311aa1b43bfd8b78bbcc8090f6bc81c200beea8", "enabled": 1 } }, "entropyd": { "100": { "checksum": "sha256:e2aa3a7bc12f543b79842c2504cb1b53b73c55c78a42e24fbead6ab2deb120a4", "enabled": 1 } }, "exim": { "100": { "checksum": "sha256:b097146234c292a604496276a05ae2cbb09bbfdbf0ac9d7948b42afddd953347", "enabled": 1 } }, "fail2ban": { "100": { "checksum": "sha256:c51270c9abffd1d173532bbce08b28480f8e261d099370268b5364c343726140", "enabled": 1 } }, "fcoe": { "100": { "checksum": "sha256:a8a2c86fcc37532b9bf516e070ff53c9e675cb5e9f154966dde6a9b5fc6d7bc5", "enabled": 1 } }, "fdo": { "100": { "checksum": "sha256:7ecdacdeaddd502ea66686e0eaa24fa19fef7202f0092a7174db2aa86eddae8e", "enabled": 1 } }, "fedoratp": { "100": { "checksum": "sha256:85ee3ad2c50bf2b0ba1fb4fe0c57850f9ffd3bb5c11d11a5b5e159ac65783c18", "enabled": 1 } }, "fetchmail": { "100": { "checksum": "sha256:56ef87427ba73e1981192f595dd2c169b322cbed91da8de818e995434e07c757", "enabled": 1 } }, "finger": { "100": { "checksum": "sha256:9aa41b846e607a383c12d0bb4a9695058f8c9b1ed4b26bb59a96855015d9ab35", "enabled": 1 } }, "firewalld": { "100": { "checksum": "sha256:9e85ce564ebec8cae461d093727c4e339a32b10ce3efa569b5c2d53c27baae59", "enabled": 1 } }, "firewallgui": { "100": { "checksum": "sha256:433e16150ab0650ca51df171c1fbe1406cfe15dfa077bade2e48a7d8514bb0ba", "enabled": 1 } }, "firstboot": { "100": { "checksum": "sha256:a16615a4aea8b12fce51f6d7715627da3c3f6c6099db8c3087e29c920d81f861", "enabled": 1 } }, "fprintd": { "100": { "checksum": "sha256:1cb012beb208aebe26b53919c54e8e902210feb7b26d52b05cbe760d273b0645", "enabled": 1 } }, "freeipmi": { "100": { "checksum": "sha256:f62cf10a14cb8de0f158fed8cc9400a8769be1ffb2f3dcdcd0e920de20d68007", "enabled": 1 } }, "freqset": { "100": { "checksum": "sha256:5c96b7146bde0acec28d73afc25b0a35fbfffe2ccd69529717eca6927dc558a7", "enabled": 1 } }, "fstools": { "100": { "checksum": "sha256:0fdc4babffbc8fa18d9445b74b14fc08b77fd6e371c96dc1f5283d6ff7c5f546", "enabled": 1 } }, "ftp": { "100": { "checksum": "sha256:dae3209382de68fdb49a72792108ec3a55c0e50abdfd38c5fae300da4c561c47", "enabled": 1 } }, "fwupd": { "100": { "checksum": "sha256:4f1c84d8ccd64b0af49edc9247ac92f49545fe4dfb24906acc133ebd7ba78bb9", "enabled": 1 } }, "games": { "100": { "checksum": "sha256:7d9b09309f694125d4ba8e1f5683baedfc2c935f0242be94531b43d842509ae9", "enabled": 1 } }, "gdomap": { "100": { "checksum": "sha256:bec0b003cf82cc2fe7e6f196bd82b929e87a827fe8d46bca5b1df2613a409d6b", "enabled": 1 } }, "geoclue": { "100": { "checksum": "sha256:3e8a90189684079f66e782c8e001c435ada771766dd1fa9ae33a977405f6ba99", "enabled": 1 } }, "getty": { "100": { "checksum": "sha256:e5ecf805b71dfe0f039f8eeefa730893c5417fe53b0e0eeba55f99670b4c4584", "enabled": 1 } }, "git": { "100": { "checksum": "sha256:69850d2342f0fcb96a719d39a3698d547c105880e6834f232a4d845a18a046f3", "enabled": 1 } }, "gitosis": { "100": { "checksum": "sha256:24935d3495375c7f6e0a0ddc091cf1f3f276ee155d07760bd7ba0c3ac115f6e8", "enabled": 1 } }, "glance": { "100": { "checksum": "sha256:fe00543de47f940ba5bf1ab84b7dc7f6a680053ef148329f2a9ec827eb47c61c", "enabled": 1 } }, "glusterd": { "100": { "checksum": "sha256:4d2befdcd924820622a0c6653772a9f5ba96aad9a3bd7b4d86daab96b233c9ea", "enabled": 1 } }, "gnome": { "100": { "checksum": "sha256:a5e2adeeda1a5b359025d956cbf20eb3773341207c72cdf7bf5001deeedf4712", "enabled": 1 } }, "gnome_remote_desktop": { "100": { "checksum": "sha256:d926eb26b34a6253b249bc1cea7696dae87887739ef77cd8cb92bdf7f6d1f371", "enabled": 1 } }, "gpg": { "100": { "checksum": "sha256:a9180cca4098d53514cd5f701f4afd40420a3664e8ffa86797ee3a8fadd9a84a", "enabled": 1 } }, "gpm": { "100": { "checksum": "sha256:2b5e83f7513a2470766cbaad46e28604ed8fcd6b640dc0350d90731a807f56ad", "enabled": 1 } }, "gpsd": { "100": { "checksum": "sha256:b64db8adeb9e8602211a0924c1b4b917d0f701df86984e73412344c9b3c2f5d9", "enabled": 1 } }, "gssproxy": { "100": { "checksum": "sha256:a43f376f2b1605cf0d45884b818db045eb37b4f037ae7af36ec5bf993676f4e5", "enabled": 1 } }, "guest": { "100": { "checksum": "sha256:262ccc7c3b237e7ebc31631d83baf2c6cbf8438cce02ce82f008fd6a3ff81a43", "enabled": 1 } }, "hddtemp": { "100": { "checksum": "sha256:fc86a156d59e31b02ce40cbda3c24b0c0baae2ed6f0cabf84cb9a3248ccd236e", "enabled": 1 } }, "hostapd": { "100": { "checksum": "sha256:ba13f04311a741f26ac34cd39688e2f4e75a546358c63c57a2d6a916dfc18315", "enabled": 1 } }, "hostname": { "100": { "checksum": "sha256:7366a48a1774b579953d8b8ae05f9ebe07e5d27984c995afe27509eacf1d79d6", "enabled": 1 } }, "hsqldb": { "100": { "checksum": "sha256:a0c31e8462c7fbbc6579270e5e35923deff56790f565157cb9c32c4eb3d6958b", "enabled": 1 } }, "hwloc": { "100": { "checksum": "sha256:fde041b834e5939945fc0d91e45940745254e128d96625ca5ab4d6e7a4379e27", "enabled": 1 } }, "hypervkvp": { "100": { "checksum": "sha256:520caa8d38f07a36d096f7ee8eb1b0610f49eafedca1721b340f5cdf21e23944", "enabled": 1 } }, "ibacm": { "100": { "checksum": "sha256:4bc0f9a40691f1abb15ce6cd277d7bab4b18720373a197c8310aa63995bcda0f", "enabled": 1 } }, "ica": { "100": { "checksum": "sha256:5daeb045f8b76fed05c1145d131f77dd556de131f777b7e259d8ec57787e7da6", "enabled": 1 } }, "icecast": { "100": { "checksum": "sha256:281ed7601c5b057802da5ffdc24a7e3308e0bb2591a263186f51e35816dffc12", "enabled": 1 } }, "iiosensorproxy": { "100": { "checksum": "sha256:05dd0967ab5bc2422ac57f2ec8a354181bf7038f0d1aabc2775c7c915bbbeaf3", "enabled": 1 } }, "inetd": { "100": { "checksum": "sha256:52f6df31e38b0c20b9fc264115ce56b4d6d8048dce5b6f250a6ca1b0675a0569", "enabled": 1 } }, "init": { "100": { "checksum": "sha256:7141bbe724e0c7f4b3dd7f335ab21267676b126f0443da9cbfa58f1372deb59c", "enabled": 1 } }, "inn": { "100": { "checksum": "sha256:78ae9628e756da216964c88e987dd91e40ec2803ff209763f628a434c8bf3b90", "enabled": 1 } }, "insights_client": { "100": { "checksum": "sha256:c6e547e12ca21f1f68ba74d667fab1bcbfc8d0461fc3a560e80160f32638ddba", "enabled": 1 } }, "iodine": { "100": { "checksum": "sha256:5c8ef0d8e730a3baca51f4c64b13df56149ef304effdc71fe0fb63f90e5a7296", "enabled": 1 } }, "iotop": { "100": { "checksum": "sha256:114de54fd32be9c32e88348ddb168291c0aaaba3a7937501f49e514585228092", "enabled": 1 } }, "ipmievd": { "100": { "checksum": "sha256:5777d2cc5e62942022a8e914847dc7436264585e41e2ce703c49c735e44add8c", "enabled": 1 } }, "ipsec": { "100": { "checksum": "sha256:ffde25d3dd0fa80b933b30cd457ea900389f84d13d4ea629f54c0be59136d4a6", "enabled": 1 } }, "iptables": { "100": { "checksum": "sha256:4c92a7102c209b0edf571a3ca3b35c597bd427d0d258c6829bfd29358448598b", "enabled": 1 } }, "irc": { "100": { "checksum": "sha256:291a822392f8a93d820aa03613bd0c058cfd1118fd6cca93a12175456eb2cf72", "enabled": 1 } }, "irqbalance": { "100": { "checksum": "sha256:2fc17d69e2cba360a0d9043585a3799f86697bf9189fcde62346b41bcf5bd9d9", "enabled": 1 } }, "iscsi": { "100": { "checksum": "sha256:a91d9d18f0502679892047e7a6023107b77412ae4b3d25f75982f766c792be58", "enabled": 1 } }, "isns": { "100": { "checksum": "sha256:5372fd0c432735b6ec28e0a5cafeebd74688912c7703616ca46b77287385c564", "enabled": 1 } }, "jabber": { "100": { "checksum": "sha256:d1916b00ece1c2b546cc04a7b6695f6459724a6bb7a92be78f399f61a613af86", "enabled": 1 } }, "jetty": { "100": { "checksum": "sha256:c78ba53bd379a65017e29895da717fa8e758494e94113b7f324d73121cb35423", "enabled": 1 } }, "jockey": { "100": { "checksum": "sha256:4cee27b59fe57c672587d220f3061417207a02bc0ff36188148f5c5f3101c1ad", "enabled": 1 } }, "journalctl": { "100": { "checksum": "sha256:27aeef17f22b9e891c4c793870f1124ec69ec7dc8f532bfec5b2a0e7319866cb", "enabled": 1 } }, "kafs": { "100": { "checksum": "sha256:602c3c40b28b653e11110e83c7f63122a75b9ad5cb7605299f6951a485284727", "enabled": 1 } }, "kdump": { "100": { "checksum": "sha256:bac33f3ca081f5f4b58fbbdbd3630aef7fb77c19368ec2cef66494cb8ef7fd90", "enabled": 1 } }, "kdumpgui": { "100": { "checksum": "sha256:7e16a86780b3b3d8a6afc5a09335cd6bf6bb39eec53815af9bbc53844655846c", "enabled": 1 } }, "keepalived": { "100": { "checksum": "sha256:5e76b1d8c27eea054956c86458f44ebc46d803035e2b3c0af67fd5feae26b9f2", "enabled": 1 } }, "kerberos": { "100": { "checksum": "sha256:65a81f3d94068c6e7cd3c1092cfbc080f853c7e02490356f7f16d40a3e8ace49", "enabled": 1 } }, "keyboardd": { "100": { "checksum": "sha256:e267709b6fc5f39679c3156b764ac82facb0b228e5b300d891f4a182662a1a5b", "enabled": 1 } }, "keystone": { "100": { "checksum": "sha256:1a7021842ad7350bc5674a455f4232b831ec8153d14635eba8930e8ade0980b7", "enabled": 1 } }, "keyutils": { "100": { "checksum": "sha256:0dc521ba68e3d21396cf5bce20427e6c0c5dfe635c743ccd4374beadd1a96cc5", "enabled": 1 } }, "kismet": { "100": { "checksum": "sha256:4398910d875745fbd61926116eddc0c7f3f6da3cf8b62fb4149d78734c344df3", "enabled": 1 } }, "kpatch": { "100": { "checksum": "sha256:36ae9280a2f7e86befeaf301f1d47087b9214bc97e79562105fcaa1246512467", "enabled": 1 } }, "ksmtuned": { "100": { "checksum": "sha256:66f6fee140f5f767906ea257b32ede5d214b8da0db4031313eb1fddb9a589d21", "enabled": 1 } }, "ktalk": { "100": { "checksum": "sha256:e783786109e8dac25e8be80e5782ec260343cf0db5b03126b85811a101bd9d0e", "enabled": 1 } }, "l2tp": { "100": { "checksum": "sha256:55f189e95389ac2f9205270f3dd53e25761d8e626989bdfb3c05b19c0086dc30", "enabled": 1 } }, "ldap": { "100": { "checksum": "sha256:cb97aa280c3bb13b6c8804e2a2bf971250db72b77ba5ece99702ab6434c3d14a", "enabled": 1 } }, "libraries": { "100": { "checksum": "sha256:bb45e05979c90858a0ed2ed0ecae78cc4e99b750145947a2940eb0493249d161", "enabled": 1 } }, "likewise": { "100": { "checksum": "sha256:f535c1e16dc83389fd5a2fa4716615c687a931e3a9ad7f65dc8607ef305f7994", "enabled": 1 } }, "linuxptp": { "100": { "checksum": "sha256:0d6f39eec4849cc1f5ca0014c29399b78317216ed0e73dfc524be3194ff1bc08", "enabled": 1 } }, "lircd": { "100": { "checksum": "sha256:cdd2ad3d9fa15b37c76e5bf0d781314f3d50897ef6cf377d70706d8a139639f7", "enabled": 1 } }, "livecd": { "100": { "checksum": "sha256:46a683c9c0693f57e0a687926c0ea32a8a176eaa168d1bf86e66463395b314cf", "enabled": 1 } }, "lldpad": { "100": { "checksum": "sha256:9fa26b1dd3ca8e52c93a1320fb66fceb0c8e83730227023d221a6a426b9dddf0", "enabled": 1 } }, "loadkeys": { "100": { "checksum": "sha256:8e876286cd84dd189d40234f691174bc5e0c214056ff0a9f454400cdafdcd6cd", "enabled": 1 } }, "locallogin": { "100": { "checksum": "sha256:f90a7ca564fcff2da3a9effa5c50e8c2a92de3e4c3831f741bc3f8cfc5be5808", "enabled": 1 } }, "lockdev": { "100": { "checksum": "sha256:d181bc52a0d2af9fe776462be417f9f28920a71f01f9894557e9bf7e128ddb5c", "enabled": 1 } }, "logadm": { "100": { "checksum": "sha256:c573637379d69beb37674b0ce9da5f57e60238a88ddad2faa1c3d456cd7df71a", "enabled": 1 } }, "logging": { "100": { "checksum": "sha256:d3494d1da872d6c128a1ab751a7ec84abb1ec73dccbfae200707935617fb8888", "enabled": 1 } }, "logrotate": { "100": { "checksum": "sha256:97dfef99f123dd0474194e78c4e8760df60d46db61f8460161bbcf4b78f7c530", "enabled": 1 } }, "logwatch": { "100": { "checksum": "sha256:c68409b41b76b32647af8f19d1f8f6ebd43430dc820e55d2b094730950a96419", "enabled": 1 } }, "lpd": { "100": { "checksum": "sha256:24594a8f8d0e864712707e8c4114f6d68dd19824045a9f3d10b8f590fc506858", "enabled": 1 } }, "lsm": { "100": { "checksum": "sha256:7df22b87a981a09999b90bb4ecea9713b595814a385fce77eb67e3c20042aa09", "enabled": 1 } }, "lttng-tools": { "100": { "checksum": "sha256:9f30b434e6c08d9a155bbc24ef6d71745d8d252c96d0ded0ccb6d158354c6546", "enabled": 1 } }, "lvm": { "100": { "checksum": "sha256:f0b2c006055d609858e59815fcd002fab6ce4a77761dfd2cc18b56219bf56fc1", "enabled": 1 } }, "mailman": { "100": { "checksum": "sha256:50901ef3e7ccc1828edc3fde37485af69afe28048ba25b7d9e3578b063d559d7", "enabled": 1 } }, "mailscanner": { "100": { "checksum": "sha256:914891424131853f76600695c57ec6283f9cc3c501c06e1427ef99f83431d1cd", "enabled": 1 } }, "man2html": { "100": { "checksum": "sha256:a7af8f4be03248a6d91c5df6c2becd22e3f45a71cecd671d44d2830fcd9f4a3b", "enabled": 1 } }, "mandb": { "100": { "checksum": "sha256:4afad481406029752a7960f97c253b41b8251018b219e16a858ccd029f9bb5f2", "enabled": 1 } }, "mcelog": { "100": { "checksum": "sha256:047b1501df33727c4257a655f1ad55aaed7da1d763782f7da45dfb1b19fa6a02", "enabled": 1 } }, "mediawiki": { "100": { "checksum": "sha256:7ce2990caacbf36b3ea6481def0630fa02e220ea9a4e2dc107251bce1368e569", "enabled": 1 } }, "memcached": { "100": { "checksum": "sha256:577518e3f4a348aaac570906781e3db46d9d53bf52d1a6f83c9f520dab8270e7", "enabled": 1 } }, "milter": { "100": { "checksum": "sha256:aa6812766cbdfad665abf5ad9633ffb945822d31714ffaa2bf4cb061f91e6b6c", "enabled": 1 } }, "minidlna": { "100": { "checksum": "sha256:f13c8b27b597e55d9e6342c756bc4d21de401fcbc4cb01e16c71c01a816c2e2b", "enabled": 1 } }, "minissdpd": { "100": { "checksum": "sha256:18649df6879516ea838eabbd19d6f7aefe2bb7da501c0b471c453d80ea4e02cb", "enabled": 1 } }, "miscfiles": { "100": { "checksum": "sha256:6f5792cde030e56b55f4b00cb53bd6d3724181c80aaaee5806e88c6a5298bb6c", "enabled": 1 } }, "mock": { "100": { "checksum": "sha256:3fdad88af11ebff86688d4db404640a1371af801e4a9b2f14a1b2abc1c76a503", "enabled": 1 } }, "modemmanager": { "100": { "checksum": "sha256:55370db1ae9a5e7ec0b4ec118f45ed61578656c6df3deb7b9e618362ddce4e29", "enabled": 1 } }, "modutils": { "100": { "checksum": "sha256:37b0e80e0d823ecd2606ece25c0b93eaea9f16da35261d79630b3e95c33a5fb4", "enabled": 1 } }, "mojomojo": { "100": { "checksum": "sha256:8ddaca2a54241adf5d42f9205cd073c7062e7692f31ebe15eb9e988e12efc9d4", "enabled": 1 } }, "mon_statd": { "100": { "checksum": "sha256:7e1bf20f24f949d86cf2cf1ecd59aef5aee94c9258d32a8fb4ee7646e4dc67e7", "enabled": 1 } }, "mongodb": { "100": { "checksum": "sha256:051c9eb43df88894edffecf5083ac5065a7051b743068b6f1866d923bca5ab95", "enabled": 1 } }, "motion": { "100": { "checksum": "sha256:154f6949eddb0f008a3b9166f736dde00bbc37df415f07c1263cd63ee609a5fd", "enabled": 1 } }, "mount": { "100": { "checksum": "sha256:acb9c37d19d0aec956cffe6f83d9593cb70bf8ad71ecf956e074b590286f489b", "enabled": 1 } }, "mozilla": { "100": { "checksum": "sha256:0a4982ad6ec48750689dc3c5d7ee946fca86c0c8b3763ee322aae33546109bed", "enabled": 1 } }, "mpd": { "100": { "checksum": "sha256:8807a444e6651ef4e67d0fed3bbc2558dc34b15e7293bfa03487d0a2c54ba905", "enabled": 1 } }, "mplayer": { "100": { "checksum": "sha256:3f999e549842f5a13f5f0d5aa4a752e74216c123f34f1797588b73d504afdc81", "enabled": 1 } }, "mptcpd": { "100": { "checksum": "sha256:f715709d773d8da3d99b4a7de877c7d7a8964d7a1342a61779edcdde2c75702d", "enabled": 1 } }, "mrtg": { "100": { "checksum": "sha256:740151127f1bcdb7699a6317eca571399e7cf58a755afd20dd141aec3035ab53", "enabled": 1 } }, "mta": { "100": { "checksum": "sha256:e04f6d7a528f9eeb7835428fe62d15043b9f4642cded56d682551e18920d04b3", "enabled": 1 } }, "munin": { "100": { "checksum": "sha256:1d2ea7b3403a2c521005ef980913298f01b2c00338366997273344590d0b218e", "enabled": 1 } }, "mysql": { "100": { "checksum": "sha256:07bc412265c70dcb7a3f85ba5265573ef76373364bbbd165ad3abb662e4de705", "enabled": 1 } }, "mythtv": { "100": { "checksum": "sha256:4cab8df79dc5a8e73ece06c418cb3ca58c74f669516b810de46b9f268d78ee8c", "enabled": 1 } }, "nagios": { "100": { "checksum": "sha256:712912fdf95fbf2adc1cb7427524d5536f0cd591bf3e77b63ef069213ae86704", "enabled": 1 } }, "namespace": { "100": { "checksum": "sha256:c032c4722dc3ec48bc9f196ae3de9ef9fd1c04554aa0913d99ade36ac0cf85af", "enabled": 1 } }, "ncftool": { "100": { "checksum": "sha256:e23fc9d15551fc1d7ba386350226d425282fcc090b9dfdcdb7065ebf3385666b", "enabled": 1 } }, "netlabel": { "100": { "checksum": "sha256:d9b7f3b8cb9b0ca564b43893085106dab4a6f6e7bee3143d778fc28218910435", "enabled": 1 } }, "netutils": { "100": { "checksum": "sha256:e828ed9b31086b0b53a5cf66cb2a580aae3e3290156b0d32c07eec5b1bd9cfe5", "enabled": 1 } }, "networkmanager": { "100": { "checksum": "sha256:bc61c1cee0f527dce7154fefb7906e6262e81d84108c7999f8db9cba7e4eb44c", "enabled": 1 } }, "ninfod": { "100": { "checksum": "sha256:4ed636e37871cc6a707bd29f6d990f4ff13d98bac2d8bf37922983d6289b3411", "enabled": 1 } }, "nis": { "100": { "checksum": "sha256:bce01b1750e64ca5f88e7d00d7d4dc88d5ebd430561f9b3cc680484b631c6c8a", "enabled": 1 } }, "nova": { "100": { "checksum": "sha256:16ebf87ef7e152147a0aa89e34161622d387348d4376f2eb91d971ef8ab476a0", "enabled": 1 } }, "nscd": { "100": { "checksum": "sha256:c7e5598679b39499dd00ac032b9a85b6e7e367bc64a0e2827cd799bb7e920a72", "enabled": 1 } }, "nsd": { "100": { "checksum": "sha256:75e019e0b2eda6179a3b39defbd43329dff4627e2c6a06c868ebee8b584332ee", "enabled": 1 } }, "nslcd": { "100": { "checksum": "sha256:69c6eaa2ef51784392d46e962c0f9b9439616dc97ba3105ff54e275a7b1e9db4", "enabled": 1 } }, "ntop": { "100": { "checksum": "sha256:cd4113e904b08b55f2a611ab02edbba284a91097a468b0d6b0d98a5dff397220", "enabled": 1 } }, "ntp": { "100": { "checksum": "sha256:5f188e0c5dabce04e39700cab0a3052cde82676b66e13b054d55f6f78f8d4b30", "enabled": 1 } }, "numad": { "100": { "checksum": "sha256:ed1193b412ae24e79c9a91531c99520b68ded7151bbf0f8cf0353d0dab713dec", "enabled": 1 } }, "nut": { "100": { "checksum": "sha256:54c04020d40b310f4b85f747395b3183859d816a6f181e4a54f5bd6f23b660df", "enabled": 1 } }, "nvme_stas": { "100": { "checksum": "sha256:1098f37f7b13bb6fa26d45bd4c54fd8cb38eddd1b4d02efaf88241f7338022c2", "enabled": 1 } }, "nx": { "100": { "checksum": "sha256:a8b41cc8424e4d9d67310c486221a9fc2701a9508104bdeb076f2ece6c94957c", "enabled": 1 } }, "obex": { "100": { "checksum": "sha256:4810ca21adc14e1f8b8f2a5b91e628038786e71ff84a2b3d701cdcafc6baf16c", "enabled": 1 } }, "oddjob": { "100": { "checksum": "sha256:d09199963de83937d060d3199ad48bf9f6753507c2748ad11d86628469f1e5b5", "enabled": 1 } }, "opafm": { "100": { "checksum": "sha256:020fbfc8bc4b762f26e66c399e2ae1f202c409d7ae1f10ef84f94c9672a1e55f", "enabled": 1 } }, "openct": { "100": { "checksum": "sha256:fd4ba492c3bc1b35ee1d9e4cf2b1fa21ac80ed31c9c76cb4503996f7c6367898", "enabled": 1 } }, "opendnssec": { "100": { "checksum": "sha256:cd5781c1d95be89112ad12c0d1341cc43b60259d4eea9b11442bc19da8ee462c", "enabled": 1 } }, "openfortivpn": { "100": { "checksum": "sha256:9a125b8b0328b3f01aea6aa395ce5117bc3a683e2b45b12e69dc6a95116ce203", "enabled": 1 } }, "openhpid": { "100": { "checksum": "sha256:65ec5e445fa5dd1471b83bccc319e9403a1c5888423442382751b1bf91c5188d", "enabled": 1 } }, "openshift": { "100": { "checksum": "sha256:07bea1192b1c51dd443db5a17d844ba3d7383abc3a3e7e979dcf0143753af074", "enabled": 1 } }, "openshift-origin": { "100": { "checksum": "sha256:88873289a3c1bd0c0c5d636b689bd8e1fe194ca528720c3191fb61b2ca024b50", "enabled": 1 } }, "opensm": { "100": { "checksum": "sha256:76a9f4d16f6b8fe2cb94b30b0c557dd4b067a4cb49502ff812b95290539062f2", "enabled": 1 } }, "openvpn": { "100": { "checksum": "sha256:ec0e41fdd7c04028cedf656a3fd60533adc778663a70734f5f2d569f5e79f53b", "enabled": 1 } }, "openvswitch": { "100": { "checksum": "sha256:1ab8ca3123bd0bbfbe2baac2e44124259401557c590cb8e906b06f28a4b9e32d", "enabled": 1 } }, "openwsman": { "100": { "checksum": "sha256:417b01a65501649535510d72e5fca7174183b5c873c1ddcb6921c052c8be3fc3", "enabled": 1 } }, "oracleasm": { "100": { "checksum": "sha256:6070bb82ad7b693e1ce70bcec6b975c4f5baec8a51e879f27e2c80ff2e076755", "enabled": 1 } }, "osad": { "100": { "checksum": "sha256:7bd10b072510204cc5f32e694b6eb1389fc20c40186fad86b76079e774023f4b", "enabled": 1 } }, "pads": { "100": { "checksum": "sha256:aeae1d20fad2ecc5e2f3ac84829becd50d70a7c41e8183c5b817e0d3f97f9484", "enabled": 1 } }, "passenger": { "100": { "checksum": "sha256:93cb4d1c0096f51325fa10c42ab72db38b5974c42b7d9fa6789482f56ce688c6", "enabled": 1 } }, "pcm": { "100": { "checksum": "sha256:4bb8d2c73a705f3051ce26155740950675bb556b11e2a7f59bb390ebbf7f9a87", "enabled": 1 } }, "pcmcia": { "100": { "checksum": "sha256:722b33d72298e2c8fa2900f1592c92673d098fcb6cd5802449086c918ca104dd", "enabled": 1 } }, "pcp": { "100": { "checksum": "sha256:a34e540529562d554136e3710fe67e8ca6e9ccee02b76dab369a448b58236f34", "enabled": 1 } }, "pcscd": { "100": { "checksum": "sha256:f50088e07d5fd52c3ad752e30852f0d3085cae9f13d2da6fd8c8141f1568ca28", "enabled": 1 } }, "pdns": { "100": { "checksum": "sha256:2d20aff28dd9cac707097c581edf455d754629927a32cc07854eb9baedb23b0a", "enabled": 1 } }, "pegasus": { "100": { "checksum": "sha256:81bedf6564a2a1e2b7ef03a9e5f646184f770202ad403e031f430e494abf8e88", "enabled": 1 } }, "permissivedomains": { "100": { "checksum": "sha256:2453bad4ace526f3cf2c60b358e95a5476692ef25da107b10f52f3af27c056d2", "enabled": 1 } }, "pesign": { "100": { "checksum": "sha256:544140731f1c09b504a6b85ec6ffbdda47c549c9858d6c2264498bfc56fd6127", "enabled": 1 } }, "pingd": { "100": { "checksum": "sha256:a4f1dd19ad8877fb1b48c431790bc350af44d3e2d0d08d59497d2a4425ce2d80", "enabled": 1 } }, "pkcs": { "100": { "checksum": "sha256:c0fe90911ac32abd43e43d9e25dd5a15c85301a1ad0ee579fd3997b9c25ad201", "enabled": 1 } }, "pki": { "100": { "checksum": "sha256:16249004873cf9214f949f73481b5476c0ab03e349542512af9b57aeb637c047", "enabled": 1 } }, "plymouthd": { "100": { "checksum": "sha256:4e33d3ebd3e3cd1fc439ca7bcf3c11086799c7af6a1e12460eb4fc5bab23b0be", "enabled": 1 } }, "podsleuth": { "100": { "checksum": "sha256:7f525e710a117aeaec78a9ab1cbbdd54a150ffe26ed5a1bc29595824fd9e1931", "enabled": 1 } }, "policykit": { "100": { "checksum": "sha256:fe85c9803d13edb28fd5a47803918af2a1e70d66b57e920fefe8e6cfdf778171", "enabled": 1 } }, "polipo": { "100": { "checksum": "sha256:8ab8b8f9438c6ae6c96323b6cf5937528d8223c38629f9bed4a2df00f4b9b686", "enabled": 1 } }, "portmap": { "100": { "checksum": "sha256:59c39477c07ba01841c150a2232e21074575ad3f93782e73b35da8f7acb121af", "enabled": 1 } }, "portreserve": { "100": { "checksum": "sha256:00aa972aa3ea3cf49a0134ebc75df0d6b7eb87ca7b31c61b3107001111abda8a", "enabled": 1 } }, "postfix": { "100": { "checksum": "sha256:68f178fd86c386a5d4df236f983239aa04eb6482304df3e9eb56340b744f5d8c", "enabled": 1 } }, "postgresql": { "100": { "checksum": "sha256:bab7ecb38d7aa40fc1de9d3f1a06e0f83c473e22e2f2391cbdc50e27ef92f7dd", "enabled": 1 } }, "postgrey": { "100": { "checksum": "sha256:8e1915326802644782f3041e65e3b1dfb6f534f2a66f4cb00ca8391f35ef2048", "enabled": 1 } }, "ppp": { "100": { "checksum": "sha256:912c8757c9d3cc46657aa135743c58836db26f4a6abf17807017c86a8512d90f", "enabled": 1 } }, "prelink": { "100": { "checksum": "sha256:0b0ad935c694badc8da9368226c53640cfade72c6828d094924029340ad5521e", "enabled": 1 } }, "prelude": { "100": { "checksum": "sha256:92c5efa0190469791dd3a105bebd81b2891596683001c6cb0d7446fc0d5125a2", "enabled": 1 } }, "privoxy": { "100": { "checksum": "sha256:0c0a1216baa5adcd33f85d92f248893c11cb55b5b068db990b2f94d9d7914932", "enabled": 1 } }, "procmail": { "100": { "checksum": "sha256:ab3962f174f9a9a125e2aad039584a83294b11d3ee56c8b477217a8193afc87d", "enabled": 1 } }, "prosody": { "100": { "checksum": "sha256:3573621c7a830f0d511041287fe8bb073eaa5e200c5b8b2e72096bf469c35f1d", "enabled": 1 } }, "psad": { "100": { "checksum": "sha256:72c77f99e96935aa093669bb095f914484b67e54ae66b4ef8ca780d2758cc6c8", "enabled": 1 } }, "ptchown": { "100": { "checksum": "sha256:318c62c2f63baefa43a61f8ef58bc168596087cd892a63afee72b1cb29cac0c7", "enabled": 1 } }, "pulseaudio": { "100": { "checksum": "sha256:3a7087d4680d1a35893caebf703937f36c1b1799dabc446b02428afc6c163c6f", "enabled": 1 } }, "puppet": { "100": { "checksum": "sha256:366806c68d03db1138cc9a3a08e0fdc4883825c20d1bed9a8e64ce7662bfb032", "enabled": 1 } }, "pwauth": { "100": { "checksum": "sha256:d23c293e39c1a025d375b714dd9319d27545c6c62b58342ef530763b84a24b16", "enabled": 1 } }, "qatlib": { "100": { "checksum": "sha256:cef00aa0bebb89745286c7db05374eb06fe9637c480d295f80f31d21be53672f", "enabled": 1 } }, "qmail": { "100": { "checksum": "sha256:5617289b46ae349fafd315cf92082d64119c6a0b5606bf79aa17fc99e34985b6", "enabled": 1 } }, "qpid": { "100": { "checksum": "sha256:1ee83ce320b057b0569c4bc9f8abdab6233eaf6fcd9c1a78975040bb2d9ddc56", "enabled": 1 } }, "quantum": { "100": { "checksum": "sha256:704e2c7d5fca4cb84e320b9359c6168037ebc87908330c0053c143c8821d3863", "enabled": 1 } }, "quota": { "100": { "checksum": "sha256:58dca8ea36c665ab74eb79cd51d54a7224514660bdfc4b666c2b8031e2877da3", "enabled": 1 } }, "rabbitmq": { "100": { "checksum": "sha256:acc51303259d5e6049265839c5ed13703baa80ae3477ae292103b4665c83dfbb", "enabled": 1 } }, "radius": { "100": { "checksum": "sha256:b53cd6e9f7493004e2b5bdf30e368fbc90b4f971db3fe56e55f31cc839969237", "enabled": 1 } }, "radvd": { "100": { "checksum": "sha256:c909b5d3b8ba285d71358dc7c9fca270e7b1586266c597a60b167ff600d10f7b", "enabled": 1 } }, "raid": { "100": { "checksum": "sha256:071d5ed18ad1c7d334bfba17d392c80a588ad00b1052ab0e21f9c53b0db95e4f", "enabled": 1 } }, "rasdaemon": { "100": { "checksum": "sha256:6d5c3ce954d1cef655ba323f1fef77bda8813815c0d991d020e0c760a4089aa5", "enabled": 1 } }, "rdisc": { "100": { "checksum": "sha256:3b10e49a0fec5da0e8998d77835464b25ced6ee9bd04a24dd861601a9aa02d2d", "enabled": 1 } }, "readahead": { "100": { "checksum": "sha256:e5b3d0b401606c60c1acdfce14c7b6b3f3363a1f58d8d91dd3be36e3df38005f", "enabled": 1 } }, "realmd": { "100": { "checksum": "sha256:b37f4de6bcd25f24eee675c18e2f1842023f0cf9a9a9a7923b64f7eaa875494f", "enabled": 1 } }, "redis": { "100": { "checksum": "sha256:b862f723b5285ad797d5a7546a17b1080690d9218a23b84bb0f7fe618169610f", "enabled": 1 } }, "remotelogin": { "100": { "checksum": "sha256:622d5ae5349b504c7ec9e472a86c5a1e1bc27606868d53d154112a63299c8dc9", "enabled": 1 } }, "restraint": { "400": { "checksum": "sha256:95d0d03fbc1d4147f02a0b3da7cd76efbdd75d1f5812cf6160e922336abbf270", "enabled": 1 } }, "rhcd": { "100": { "checksum": "sha256:fcd8c8415a58d937e726318252127c28e69af2819be01b977b4dda03994834a0", "enabled": 1 } }, "rhcs": { "100": { "checksum": "sha256:1e8298e1e453be52e1a86e391b827f2c71729c066f7d6c63abe7843e28647207", "enabled": 1 } }, "rhev": { "100": { "checksum": "sha256:eaa7e5657902e356ab7700656ff4c185eaf8df675d0a9a3f9ea1a4d8d8fb577d", "enabled": 1 } }, "rhgb": { "100": { "checksum": "sha256:5271469985f35fb7911a72e9e705bc35971294386838f1f11033657874821cea", "enabled": 1 } }, "rhnsd": { "100": { "checksum": "sha256:2a502ce16b7c56df48d0798ed7ca60370d0993ee425a542bf0621b51d4caf2d4", "enabled": 1 } }, "rhsmcertd": { "100": { "checksum": "sha256:80b01c4c3ddf665dd2fd5391b6069e9b868573f7c7a9d4421bc5bfeaadc71a22", "enabled": 1 } }, "rhts": { "400": { "checksum": "sha256:39f15722a115b11064c20b8bc2758e8fe06531a8f923ea00a3e7079a0554e3d6", "enabled": 1 } }, "ricci": { "100": { "checksum": "sha256:dd355c8a0a048baee894e0951365b525bbf5d7cce5907c9919bb0fffb1876661", "enabled": 1 } }, "rkhunter": { "100": { "checksum": "sha256:95ecb10dfe02e0384b0f220c46a7d0c85166a4e2cca0bb55654420b73611a646", "enabled": 1 } }, "rlogin": { "100": { "checksum": "sha256:d8594a73f246468d42569485bb148cbf1b582779074735f6da46e92f763984e1", "enabled": 1 } }, "rngd": { "100": { "checksum": "sha256:50b57a4776d2c6de8b1698bd31ecbfcc98d90d4a9bd2c95672b114c9637ce940", "enabled": 1 } }, "roundup": { "100": { "checksum": "sha256:54c5e85363d326f78c5896bf87c36fca9bc28acee3bd9da4d6d22f090dc9f7c6", "enabled": 1 } }, "rpc": { "100": { "checksum": "sha256:404f69ae03180ddce44c15efc636e5d7ff34b2108cb2c09624075be42a0545b1", "enabled": 1 } }, "rpcbind": { "100": { "checksum": "sha256:f5c703b4107b2616b1adf105b5c592c79f03df1f3a24f31453a242cdec50ea97", "enabled": 1 } }, "rpm": { "100": { "checksum": "sha256:bb52185531861597a0b44bca75497c08ac4c07e10ec63505c333bde800176fb7", "enabled": 1 } }, "rrdcached": { "100": { "checksum": "sha256:2cb11286a0afa94eccbd33de5a32b16d9e178e41ac1fec9dcd801b77dd257051", "enabled": 1 } }, "rshd": { "100": { "checksum": "sha256:0052600f999e54f224d1210c5920b42227cd98ce8db24155b0cfecdf4f6c6474", "enabled": 1 } }, "rshim": { "100": { "checksum": "sha256:7825528bbcdb081ead1d0d2da0d5156e430f93d003f58dcebbb8997c906f0dec", "enabled": 1 } }, "rssh": { "100": { "checksum": "sha256:cde9bfc64b345835e9f8012426cf218bad9972c70e22e0ef6d6a90cc5fa7163c", "enabled": 1 } }, "rsync": { "100": { "checksum": "sha256:1295d1a5434aa1df7fd0f0e2a9b2b045b0ec51bd8619d528349530b322eddf24", "enabled": 1 } }, "rtas": { "100": { "checksum": "sha256:8c5c3a495b42ed14f56e360dd9eaaa0df2426f2c93676880f129bab95e247338", "enabled": 1 } }, "rtkit": { "100": { "checksum": "sha256:7b0c8b0fd23723aed98bc581f2ead99da87893c6ba042ec19ca731e1d802eab6", "enabled": 1 } }, "rwho": { "100": { "checksum": "sha256:fe02b4fb75b8a75ac77b53ae13939199422543f68e71829fd1530aa34c8715cc", "enabled": 1 } }, "samba": { "100": { "checksum": "sha256:b0af5e99697f8e864538813653ea879ea50b6a9269424ca1deaa95ec1a4a7778", "enabled": 1 } }, "sambagui": { "100": { "checksum": "sha256:cde09885f6d657107c517f09f01933816199f0694799fe3df4ac281b411edbb9", "enabled": 1 } }, "sandboxX": { "100": { "checksum": "sha256:ffd79700483dc635b0396df2f9ca75062aff985e7874c2014bca95e80b08bd5d", "enabled": 1 } }, "sanlock": { "100": { "checksum": "sha256:fcd9cea4d61e3597a5e5cb1bd36efc228a21e5b8ee79c68b44e8a93aae834fb6", "enabled": 1 } }, "sap": { "100": { "checksum": "sha256:4cffd87a27c2b48cc59f596a42e860ea52acdfcbff0c2a1e4aeafa3338b29e6e", "enabled": 1 } }, "sasl": { "100": { "checksum": "sha256:ed40cd297b033fd86770692e7d3ab76598684f521f079475d8b44953116291cf", "enabled": 1 } }, "sbd": { "100": { "checksum": "sha256:8c40daa2d8288e7535e9db7724d16a34add080c65fab32bdf5e10cdace95a063", "enabled": 1 } }, "sblim": { "100": { "checksum": "sha256:73b934662023f728bfeb5cac500547e15e389bad3575a3750bf05901d4396531", "enabled": 1 } }, "screen": { "100": { "checksum": "sha256:8dcba8c325fdb2c56f7c7c77362e287469484b19f0b491dd7bc0abc6e033378b", "enabled": 1 } }, "secadm": { "100": { "checksum": "sha256:2d9e07e5ee506dbf6cf976e257210fdb14222c1e437b1f7dceb1548fa399beba", "enabled": 1 } }, "sectoolm": { "100": { "checksum": "sha256:e6c4cabc853877f459381f06d4d47660ce9360a188ab87bb825f1f1eaed39d7d", "enabled": 1 } }, "selinuxutil": { "100": { "checksum": "sha256:38da361f759758e151b7f8612954c91c0fde09615e1c82242c80706231e16355", "enabled": 1 } }, "sendmail": { "100": { "checksum": "sha256:286d18507d80fcfa7cb83ed21321df23c3a29001777890137c1b52c113cd8489", "enabled": 1 } }, "sensord": { "100": { "checksum": "sha256:6681fdfffe9ba81fa3712f665b98feb68f9331f8f282faf19a4ae5e70b02bd63", "enabled": 1 } }, "setrans": { "100": { "checksum": "sha256:69a3bce8a9e35afa4c3b4c7e2f4ba604491313a688d36e3158f96e90343ea370", "enabled": 1 } }, "setroubleshoot": { "100": { "checksum": "sha256:c0d9c0c6cea100697cbc684f2ab74ec66d7c121ca88200752585aeaf4c2b2596", "enabled": 1 } }, "seunshare": { "100": { "checksum": "sha256:219912d16220957b98bbe694f06c3509d2f884ff8418da1eeeda43f4dce81dc9", "enabled": 1 } }, "shorewall": { "100": { "checksum": "sha256:d0dd25c1a5b77a28fcf61f4796c479a026a8e2ec944f5e4d1492cd552aa9ca02", "enabled": 1 } }, "slocate": { "100": { "checksum": "sha256:8d149bca3982ac8c0f7376571d1f0614f05eb2fe7465951eefb17153f642ee40", "enabled": 1 } }, "slpd": { "100": { "checksum": "sha256:6798797c2dda6b8b24aada5a4bbc4f8775faca22435ed1de39193457043f8d8b", "enabled": 1 } }, "smartmon": { "100": { "checksum": "sha256:9de88cb8f60049e56dc5ead706a324189ff276929ccc3bf4ba7d06e96c0d8414", "enabled": 1 } }, "smokeping": { "100": { "checksum": "sha256:29560964ba6e3cb5fc392417ff4aa3e50a5f15c2bc21bdf37088c87868ffeec7", "enabled": 1 } }, "smoltclient": { "100": { "checksum": "sha256:0e47cf6e775a1e2ecb506e444731e086d057b7f2c2aeb6bce7b636372de88d95", "enabled": 1 } }, "snapper": { "100": { "checksum": "sha256:e264d81b3e1173c98c1efc90529d7f9c2c9047b843e29a4fb807d0b1b96b4685", "enabled": 1 } }, "snmp": { "100": { "checksum": "sha256:ec6638b69ec04510a807a5d2c97568a670f80a7999ee711e5e1874136923718e", "enabled": 1 } }, "snort": { "100": { "checksum": "sha256:67c1751c124c249fa15c36f948b60a76db3729f819678cecba569c79abec3af2", "enabled": 1 } }, "sosreport": { "100": { "checksum": "sha256:345531e273b1a0c0d6e02a1898d60759007d5126ec2016eea320832c33ac8211", "enabled": 1 } }, "soundserver": { "100": { "checksum": "sha256:932d162b8365f213b97eb6a7e62d20b8c30bfe1863b6534462da96077d6347fc", "enabled": 1 } }, "spamassassin": { "100": { "checksum": "sha256:fbceb12b459776c22867efa6bf2d07657e058c4a80d6f20e724086ec58e600b0", "enabled": 1 } }, "speech-dispatcher": { "100": { "checksum": "sha256:092593b379ff3e7b3d15b0dfdd2adfde73fb76401d1b6b6e79a4efccbf03f0c9", "enabled": 1 } }, "squid": { "100": { "checksum": "sha256:4374a853cb79bd24d05f65585ad1a74a7b9da4d3454ca233da351e009e75ed79", "enabled": 1 } }, "ssh": { "100": { "checksum": "sha256:493ee98f9cda10ca995eb5600d2940556d70cb3f544501c6bd1968745cc234fa", "enabled": 1 } }, "sslh": { "100": { "checksum": "sha256:5dc03f6d33a6caa7c2fb5ede0f8ded4ecb8eb244443614189cb5909521cab539", "enabled": 1 } }, "sssd": { "100": { "checksum": "sha256:4ee157395107db3b146239f14c9868f6e9d71199520e5d2e473e5e836d0635d1", "enabled": 1 } }, "staff": { "100": { "checksum": "sha256:6a89d52653ce210b1dfb2d2b10e184cacb18976d238fc92b349c40661fdcbf47", "enabled": 1 } }, "stalld": { "100": { "checksum": "sha256:14aa28921885421f92fb51920ca0d8fde122881ac9e28d024c80a7b234b82dd3", "enabled": 1 } }, "stapserver": { "100": { "checksum": "sha256:1610ffc61426c0e791bd7a6a2628935edf3d66fc7818fb42c314044a26ac5d2d", "enabled": 1 } }, "stratisd": { "100": { "checksum": "sha256:5025b28ad7090d4802078c92ba257b46909bbb74fcd6cae0f2a5d65978a5ff8e", "enabled": 1 } }, "stunnel": { "100": { "checksum": "sha256:67b1c960e0cd11cdf1446615cf633700585e69f41c4039cfb183d69527648f1e", "enabled": 1 } }, "su": { "100": { "checksum": "sha256:60850779bd04d036ea792653d4e794f0e865936648c03030515d8125cf412b4b", "enabled": 1 } }, "sudo": { "100": { "checksum": "sha256:0965d3f30b2813a4364e9ce8da0f8916579319ed9507640383c3bf195ae1c00e", "enabled": 1 } }, "svnserve": { "100": { "checksum": "sha256:d589d2e8bddbcea03272bb589cffde776ccb410ac685e5071896fa8e59660c7f", "enabled": 1 } }, "swift": { "100": { "checksum": "sha256:407ddd533381c1f822ec8471e639404c4b926e7e635b3d25d4ba1403a5c517b3", "enabled": 1 } }, "sysadm": { "100": { "checksum": "sha256:1d310d8df59d431579fa3568b7ac14f853223aadd4087fe4c57a1add6606987d", "enabled": 1 } }, "sysadm_secadm": { "100": { "checksum": "sha256:6a0a3b3c4028f304beeacb7d06c62975a2bfb13d67a6a4fe6b95f6d6eeaa74b7", "enabled": 1 } }, "sysnetwork": { "100": { "checksum": "sha256:56b999cfaf7defa2c5cdd5c0c16e7c9c1de940dcc528778ff752fca51f95b38a", "enabled": 1 } }, "sysstat": { "100": { "checksum": "sha256:358b9c77d350c4bc17cf773e99f81d5267f8c41d47402234be37dd9f31655ad0", "enabled": 1 } }, "systemd": { "100": { "checksum": "sha256:124460467ec0f3f7443a67583ef4613154b2b92615e0d779a3ec8060c5a8ce4a", "enabled": 1 } }, "tangd": { "100": { "checksum": "sha256:a2e2a5d807eabe30a09085dcec38d530781c1f02e840a6ed02ccc32bd2738e60", "enabled": 1 } }, "targetd": { "100": { "checksum": "sha256:4a9e67698b4759a406188562c375ecd847f205c38873fb895c3c20abc8bac869", "enabled": 1 } }, "tcpd": { "100": { "checksum": "sha256:b69263d84bf511825be889b667401c30964c7d30d1147d577ae294cd7f05aa01", "enabled": 1 } }, "tcsd": { "100": { "checksum": "sha256:fa40e623b09fb5f61c2c2113a8ed0d17a4a5d105b9f22389ce088a766c0042c0", "enabled": 1 } }, "telepathy": { "100": { "checksum": "sha256:d98021b3bbdd6255c2009e5adafbf41faf3370434396cf2e928e7df693dd50b6", "enabled": 1 } }, "telnet": { "100": { "checksum": "sha256:abf90c09c67553f2b04cab7461138bfd8689752a752a79e36e248f209eae73e6", "enabled": 1 } }, "tftp": { "100": { "checksum": "sha256:f5dcde99c32caa66ea23a9be3b2d6d95550eff7db62b11733f16f1f5eb01bce2", "enabled": 1 } }, "tgtd": { "100": { "checksum": "sha256:6a96f6d7abaa0e1951832347e924c5aff5c66194da3881ec4855e1401a5ce6d2", "enabled": 1 } }, "thin": { "100": { "checksum": "sha256:b0c1055c6a8419035ee9a9a45de2bcdd7183a906bb40d98bb2be80b882a60c51", "enabled": 1 } }, "thumb": { "100": { "checksum": "sha256:928596446716a5d1ef1f8958d150959e59a6258975ebf453c2d5a2747e6d59fd", "enabled": 1 } }, "tlp": { "100": { "checksum": "sha256:540f0a0339feec1e11d145c6ba99f700750451a2fad43190e256e590c3835950", "enabled": 1 } }, "tmpreaper": { "100": { "checksum": "sha256:4370bc58da3c779088151878e7291e03fcd2ed706425b4f2eeeed7097f8adcad", "enabled": 1 } }, "tomcat": { "100": { "checksum": "sha256:98e40c398d94d82eea102055d52f59f77581955b0e463dd95386b4559a4cc852", "enabled": 1 } }, "tor": { "100": { "checksum": "sha256:b309ee2c391d91ace45662d2152870961635ca8968487d0f2570508e160dbd9b", "enabled": 1 } }, "tuned": { "100": { "checksum": "sha256:90bcbb5cad47c032a02390e0c4b22ff79f93fbfd2eb4a385441a87532d6aa632", "enabled": 1 } }, "tvtime": { "100": { "checksum": "sha256:a4438b3063cc1760c663e35a7be70a7dcd099de505ce1562e02090ac6b635787", "enabled": 1 } }, "udev": { "100": { "checksum": "sha256:f308ba2a718171dc871e2586f35a90cb1bbc37decf13e714af72b38cc44505a2", "enabled": 1 } }, "ulogd": { "100": { "checksum": "sha256:05aa3f423e0505dfef9b5e311b0f0a71d93159c7f6440979d3c00a5d9d22ef4c", "enabled": 1 } }, "uml": { "100": { "checksum": "sha256:e0257dd4e250c5aa0baa8151d48942735949bd3018318c11c28a4a5df2aacb90", "enabled": 1 } }, "unconfined": { "100": { "checksum": "sha256:26c4304a314b04f6269b45b4285b449caad98d22de7b3017ef648212c6674da1", "enabled": 1 } }, "unconfineduser": { "100": { "checksum": "sha256:af2778eed6f3cc25c5390e4b1a892995c37a83f5952f51820007b8cd387833b4", "enabled": 1 } }, "unlabelednet": { "100": { "checksum": "sha256:c1de2adfab2cd577a0cc9f2c1a50cd6321324d31b9c3f81be9231f85d4acbbd3", "enabled": 1 } }, "unprivuser": { "100": { "checksum": "sha256:dc1ddbf10a94d55c476ff343c862c24dc649d82190a16a0fd924c527743a667d", "enabled": 1 } }, "updfstab": { "100": { "checksum": "sha256:11eb931e80afefafd0f1c41ea19add0c746170447c76dd629146c43bd9e98067", "enabled": 1 } }, "usbmodules": { "100": { "checksum": "sha256:3a8334855e0cfe67e2c29198baacf61ea0e397d5da30f60fb844044cd105557e", "enabled": 1 } }, "usbmuxd": { "100": { "checksum": "sha256:e883ad3abf2e2792abf4122f99db904139b2f1833b7d99c45b34a7c8244d7c66", "enabled": 1 } }, "userdomain": { "100": { "checksum": "sha256:b6cc74d18c0d17741092aaa8e830f160021ad38bdc38880e3b1488ef30f59a49", "enabled": 1 } }, "userhelper": { "100": { "checksum": "sha256:0a80be6b1635de7593f46421219be5a6ccb9c22973c60b74436ba16884d789fc", "enabled": 1 } }, "usermanage": { "100": { "checksum": "sha256:c56dfa32f163d476dd5761e762ea743189a97636acaf2b4eefcf88812e624877", "enabled": 1 } }, "usernetctl": { "100": { "checksum": "sha256:9c0430aca661392960008075f0a1a934f0c05239976e73c4f0408e9cb5e279ca", "enabled": 1 } }, "uucp": { "100": { "checksum": "sha256:527db920542a250b24e24cb8c6bd86502c3be7e270d11f43d4ecb1a6632f8421", "enabled": 1 } }, "uuidd": { "100": { "checksum": "sha256:b3df65ab2d82cbb05e872f00f46156cfbae937a3a6f850d05f43f4866b772af8", "enabled": 1 } }, "varnishd": { "100": { "checksum": "sha256:9174c3c68bf6b4c2b2b9be7e2a0d985ce6edfdec0dc4c225b427b986449d3e7a", "enabled": 1 } }, "vdagent": { "100": { "checksum": "sha256:52636260af5ae220698d505c6a90a33a71f5878a2cea3a303033c896f000b732", "enabled": 1 } }, "vhostmd": { "100": { "checksum": "sha256:df2d64343b871c840603e82f5fc1f7928c0345b314450e536f9ef66becf131ad", "enabled": 1 } }, "virt": { "100": { "checksum": "sha256:c209f246ed99c66defe424c0f1bd13e970df31d84d2f4e4a2a79ef1c00c3bbe2", "enabled": 1 } }, "virt_supplementary": { "100": { "checksum": "sha256:e8147b1556404b4a2ce521cc4c22e28d4ee42045ea2480043fc8bd2d5a593207", "enabled": 1 } }, "vlock": { "100": { "checksum": "sha256:54a1b28cebc1928622edc4f1f1762a5f2d57783be153052d592f2d10dd9b24a1", "enabled": 1 } }, "vmtools": { "100": { "checksum": "sha256:323b80c9ba1efb9852725d635a5aa442ae2bbd4ec74360cf75853616401dd4f9", "enabled": 1 } }, "vmware": { "100": { "checksum": "sha256:b68b94c1c4662f7005faac011c887619dbb76d69d3d2b63f354d48a144d0d63c", "enabled": 1 } }, "vnstatd": { "100": { "checksum": "sha256:c544da94ee903d411f7ab352b3fe286fa0a1b0fb8490d508565df66d5501b852", "enabled": 1 } }, "vpn": { "100": { "checksum": "sha256:035f214c319dcee2ce4e9751fa9cf1343d81d569409f3a56faf79e506e445504", "enabled": 1 } }, "w3c": { "100": { "checksum": "sha256:d6c6c226dcc92d9a668f42b86dde398d2d875b479915af656734357e20ed0e96", "enabled": 1 } }, "watchdog": { "100": { "checksum": "sha256:a237aeaf94bed19de41ef5289be96ce5bf85dce6f2fc4f9d36d6185b4e59520a", "enabled": 1 } }, "wdmd": { "100": { "checksum": "sha256:1012fdefa5dcf2e540f536de5da79df0ba715ccc2a86a414e2e846675d32495f", "enabled": 1 } }, "webadm": { "100": { "checksum": "sha256:f374fcd23ed5f30ce1159470af1e1254e0352a6ad7cf261287de8ec51f2b2626", "enabled": 1 } }, "webalizer": { "100": { "checksum": "sha256:b66dfdfd23ed304d5f416bb0d8de4565d28cd732253fc2f9e4b4bac59515790f", "enabled": 1 } }, "wine": { "100": { "checksum": "sha256:a4a3be2a0feea4de24ca26c0eed58c7025304d810ee9d4a3e1b7e964a8d28c6d", "enabled": 1 } }, "wireguard": { "100": { "checksum": "sha256:3084b8fd20df45b718cb6779240f4292a0437f536bde7db3616c40b1500ba0f6", "enabled": 1 } }, "wireshark": { "100": { "checksum": "sha256:3ec8e55883a8b73de513f533ed80b1b665e3ed650fdf2b471e5565e1cc3034d9", "enabled": 1 } }, "xen": { "100": { "checksum": "sha256:adabf99dccce9182074a23fb8f05ac537899170fbbfe022f39cb520275339186", "enabled": 1 } }, "xguest": { "100": { "checksum": "sha256:8a7f7d9514fa5257088e451dcf96c6c221a7b6499cae9ac15a688cbe2e84def9", "enabled": 1 } }, "xserver": { "100": { "checksum": "sha256:ca8b72b391320d267580cebf2ad9ef811a781b34647e1c63e525e982911cf956", "enabled": 1 } }, "zabbix": { "100": { "checksum": "sha256:ed9800f41440cf137baa7864acdd8a503b7e051cf023d252b87daeb5357be889", "enabled": 1 } }, "zarafa": { "100": { "checksum": "sha256:c3a9d1c95f377fafb9e94a84bec89645a1053d23b557b143367992eae9993596", "enabled": 1 } }, "zebra": { "100": { "checksum": "sha256:77d57a56a3196f09ea5eb808f8c3b4cef333b241e9007b91779ba78a857e5cfa", "enabled": 1 } }, "zoneminder": { "100": { "checksum": "sha256:675bdf525f863eaf17fe6c25507cf2dc3c4ab05d1e4b0d3477573d34980a5cd2", "enabled": 1 } }, "zosremote": { "100": { "checksum": "sha256:79bb5490f9f16e8caff67fc87a834bf6f11647cfed990c421616eb1adaac5c70", "enabled": 1 } } }, "selinux_priorities": true }, "changed": false } TASK [fedora.linux_system_roles.selinux : Load SELinux modules] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:115 Saturday 11 January 2025 11:29:45 -0500 (0:00:03.061) 0:01:10.978 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_modules is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Restore SELinux labels on filesystem tree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:128 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.061) 0:01:11.040 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Restore SELinux labels on filesystem tree in check mode] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:136 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.098) 0:01:11.138 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:153 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.070) 0:01:11.208 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:157 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.040) 0:01:11.248 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:166 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.036) 0:01:11.285 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:175 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.035) 0:01:11.321 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:182 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.035) 0:01:11.356 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml for managed-node2 => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:14 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.126) 0:01:11.482 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_spec": { "state": "created" }, "__podman_kube_str": "apiVersion: v1\nkind: Pod\nmetadata:\n labels:\n app: test\n io.containers.autoupdate: registry\n name: bogus\nspec:\n containers:\n - name: bogus\n image: >-\n quay.io/linux-system-roles/this_is_a_bogus_image:latest\n" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:21 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.055) 0:01:11.538 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_continue_if_pull_fails": true, "__podman_kube": { "apiVersion": "v1", "kind": "Pod", "metadata": { "labels": { "app": "test", "io.containers.autoupdate": "registry" }, "name": "bogus" }, "spec": { "containers": [ { "image": "quay.io/linux-system-roles/this_is_a_bogus_image:latest", "name": "bogus" } ] } }, "__podman_kube_file": "", "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:33 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.054) 0:01:11.592 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_name": "bogus", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:38 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.041) 0:01:11.634 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.151) 0:01:11.785 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.090) 0:01:11.875 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 11 January 2025 11:29:46 -0500 (0:00:00.070) 0:01:11.945 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Saturday 11 January 2025 11:29:47 -0500 (0:00:00.066) 0:01:12.012 ****** ok: [managed-node2] => { "changed": false, "stat": { "atime": 1736612945.6352425, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1736612940.9742577, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 9160785, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "1643853349", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Saturday 11 January 2025 11:29:47 -0500 (0:00:00.404) 0:01:12.416 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Saturday 11 January 2025 11:29:47 -0500 (0:00:00.048) 0:01:12.465 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Saturday 11 January 2025 11:29:47 -0500 (0:00:00.040) 0:01:12.505 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Saturday 11 January 2025 11:29:47 -0500 (0:00:00.040) 0:01:12.546 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Saturday 11 January 2025 11:29:47 -0500 (0:00:00.041) 0:01:12.587 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Saturday 11 January 2025 11:29:47 -0500 (0:00:00.041) 0:01:12.628 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Saturday 11 January 2025 11:29:47 -0500 (0:00:00.043) 0:01:12.672 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Saturday 11 January 2025 11:29:47 -0500 (0:00:00.112) 0:01:12.784 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if no kube spec is given] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:43 Saturday 11 January 2025 11:29:47 -0500 (0:00:00.045) 0:01:12.830 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_kube", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:52 Saturday 11 January 2025 11:29:47 -0500 (0:00:00.063) 0:01:12.893 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_activate_systemd_unit": false, "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:60 Saturday 11 January 2025 11:29:47 -0500 (0:00:00.068) 0:01:12.961 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_path": "/etc/containers/ansible-kubernetes.d" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:64 Saturday 11 January 2025 11:29:48 -0500 (0:00:00.041) 0:01:13.003 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_file": "/etc/containers/ansible-kubernetes.d/bogus.yml" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:68 Saturday 11 January 2025 11:29:48 -0500 (0:00:00.043) 0:01:13.047 ****** ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Get service name using systemd-escape] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:75 Saturday 11 January 2025 11:29:48 -0500 (0:00:00.049) 0:01:13.096 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_activate_systemd_unit | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Cleanup containers and services] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:83 Saturday 11 January 2025 11:29:48 -0500 (0:00:00.063) 0:01:13.160 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update containers and services] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:87 Saturday 11 January 2025 11:29:48 -0500 (0:00:00.047) 0:01:13.208 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:2 Saturday 11 January 2025 11:29:48 -0500 (0:00:00.096) 0:01:13.305 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 11 January 2025 11:29:48 -0500 (0:00:00.118) 0:01:13.423 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 11 January 2025 11:29:48 -0500 (0:00:00.036) 0:01:13.460 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 11 January 2025 11:29:48 -0500 (0:00:00.037) 0:01:13.497 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the host mount volumes] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:7 Saturday 11 January 2025 11:29:48 -0500 (0:00:00.043) 0:01:13.541 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'volumes' in __podman_kube['spec']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:18 Saturday 11 January 2025 11:29:48 -0500 (0:00:00.044) 0:01:13.585 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_volumes | d([]) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:29 Saturday 11 January 2025 11:29:48 -0500 (0:00:00.050) 0:01:13.635 ****** ok: [managed-node2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Check the kubernetes yaml file] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:53 Saturday 11 January 2025 11:29:49 -0500 (0:00:00.868) 0:01:14.504 ****** ok: [managed-node2] => { "changed": false, "failed_when_result": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Ensure the kubernetes directory is present] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:61 Saturday 11 January 2025 11:29:49 -0500 (0:00:00.388) 0:01:14.893 ****** ok: [managed-node2] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/ansible-kubernetes.d", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 24, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure kubernetes yaml files are present] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:70 Saturday 11 January 2025 11:29:50 -0500 (0:00:00.408) 0:01:15.301 ****** changed: [managed-node2] => { "changed": true, "checksum": "f8266a972ed3be7e204d2a67883fe3a22b8dbf18", "dest": "/etc/containers/ansible-kubernetes.d/bogus.yml", "gid": 0, "group": "root", "md5sum": "22799c1e99f8fb14db15efc7dbc7ba4c", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 237, "src": "/root/.ansible/tmp/ansible-tmp-1736612990.3622415-9267-128608785517959/.source.yml", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Update containers/pods] *************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:80 Saturday 11 January 2025 11:29:51 -0500 (0:00:00.757) 0:01:16.059 ****** fatal: [managed-node2]: FAILED! => { "changed": false } MSG: Output: Error=Trying to pull quay.io/linux-system-roles/this_is_a_bogus_image:latest... Error: initializing source docker://quay.io/linux-system-roles/this_is_a_bogus_image:latest: reading manifest latest in quay.io/linux-system-roles/this_is_a_bogus_image: unauthorized: access to the requested resource is not authorized TASK [Verify image not pulled] ************************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:262 Saturday 11 January 2025 11:29:52 -0500 (0:00:01.819) 0:01:17.878 ****** ok: [managed-node2] => { "changed": false } MSG: All assertions passed TASK [Cleanup] ***************************************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:269 Saturday 11 January 2025 11:29:53 -0500 (0:00:00.151) 0:01:18.030 ****** included: fedora.linux_system_roles.podman for managed-node2 => (item=nopull) included: fedora.linux_system_roles.podman for managed-node2 => (item=bogus) TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Saturday 11 January 2025 11:29:53 -0500 (0:00:00.336) 0:01:18.366 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Saturday 11 January 2025 11:29:53 -0500 (0:00:00.075) 0:01:18.441 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Saturday 11 January 2025 11:29:53 -0500 (0:00:00.057) 0:01:18.499 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Saturday 11 January 2025 11:29:53 -0500 (0:00:00.041) 0:01:18.540 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Saturday 11 January 2025 11:29:53 -0500 (0:00:00.039) 0:01:18.579 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Saturday 11 January 2025 11:29:53 -0500 (0:00:00.038) 0:01:18.618 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Saturday 11 January 2025 11:29:53 -0500 (0:00:00.039) 0:01:18.657 ****** [WARNING]: TASK: fedora.linux_system_roles.podman : Set platform/version specific variables: The loop variable 'item' is already in use. You should set the `loop_var` value in the `loop_control` option for the task to something else to avoid variable collisions and unexpected behavior. ok: [managed-node2] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed-node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node2] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } ok: [managed-node2] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Saturday 11 January 2025 11:29:53 -0500 (0:00:00.118) 0:01:18.776 ****** ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Saturday 11 January 2025 11:29:54 -0500 (0:00:01.012) 0:01:19.788 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Saturday 11 January 2025 11:29:54 -0500 (0:00:00.037) 0:01:19.825 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages))", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Saturday 11 January 2025 11:29:54 -0500 (0:00:00.043) 0:01:19.869 ****** skipping: [managed-node2] => { "false_condition": "__podman_is_transactional | d(false)" } TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Saturday 11 January 2025 11:29:54 -0500 (0:00:00.036) 0:01:19.906 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Saturday 11 January 2025 11:29:54 -0500 (0:00:00.042) 0:01:19.948 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Saturday 11 January 2025 11:29:55 -0500 (0:00:00.045) 0:01:19.994 ****** ok: [managed-node2] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.023521", "end": "2025-01-11 11:29:55.332394", "rc": 0, "start": "2025-01-11 11:29:55.308873" } STDOUT: podman version 5.3.1 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Saturday 11 January 2025 11:29:55 -0500 (0:00:00.418) 0:01:20.413 ****** ok: [managed-node2] => { "ansible_facts": { "podman_version": "5.3.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Saturday 11 January 2025 11:29:55 -0500 (0:00:00.046) 0:01:20.459 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Saturday 11 January 2025 11:29:55 -0500 (0:00:00.043) 0:01:20.502 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "(podman_quadlet_specs | length > 0) or (podman_secrets | length > 0)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Saturday 11 January 2025 11:29:55 -0500 (0:00:00.039) 0:01:20.542 ****** META: end_host conditional evaluated to False, continuing execution for managed-node2 skipping: [managed-node2] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node2" } MSG: end_host conditional evaluated to false, continuing execution for managed-node2 TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Saturday 11 January 2025 11:29:55 -0500 (0:00:00.108) 0:01:20.650 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__has_type_pod or __has_pod_file_ext or __has_pod_file_src_ext or __has_pod_template_src_ext or __has_pod_template_src_ext_j2", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:96 Saturday 11 January 2025 11:29:55 -0500 (0:00:00.063) 0:01:20.714 ****** META: end_host conditional evaluated to False, continuing execution for managed-node2 skipping: [managed-node2] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node2" } MSG: end_host conditional evaluated to false, continuing execution for managed-node2 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:109 Saturday 11 January 2025 11:29:55 -0500 (0:00:00.078) 0:01:20.792 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 11 January 2025 11:29:55 -0500 (0:00:00.095) 0:01:20.888 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 11 January 2025 11:29:55 -0500 (0:00:00.049) 0:01:20.938 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 11 January 2025 11:29:55 -0500 (0:00:00.044) 0:01:20.982 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Saturday 11 January 2025 11:29:56 -0500 (0:00:00.056) 0:01:21.039 ****** ok: [managed-node2] => { "changed": false, "stat": { "atime": 1736612945.6352425, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1736612940.9742577, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 9160785, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "1643853349", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Saturday 11 January 2025 11:29:56 -0500 (0:00:00.435) 0:01:21.474 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Saturday 11 January 2025 11:29:56 -0500 (0:00:00.069) 0:01:21.544 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Saturday 11 January 2025 11:29:56 -0500 (0:00:00.112) 0:01:21.656 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Saturday 11 January 2025 11:29:56 -0500 (0:00:00.048) 0:01:21.705 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Saturday 11 January 2025 11:29:56 -0500 (0:00:00.042) 0:01:21.747 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Saturday 11 January 2025 11:29:56 -0500 (0:00:00.039) 0:01:21.787 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Saturday 11 January 2025 11:29:56 -0500 (0:00:00.038) 0:01:21.826 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Saturday 11 January 2025 11:29:56 -0500 (0:00:00.039) 0:01:21.866 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:115 Saturday 11 January 2025 11:29:56 -0500 (0:00:00.040) 0:01:21.907 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:124 Saturday 11 January 2025 11:29:56 -0500 (0:00:00.063) 0:01:21.970 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.095) 0:01:22.065 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.050) 0:01:22.116 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:127 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.043) 0:01:22.159 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.121) 0:01:22.281 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.040) 0:01:22.321 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:130 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.038) 0:01:22.360 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.099) 0:01:22.460 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.055) 0:01:22.515 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:133 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.039) 0:01:22.554 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.075) 0:01:22.630 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.049) 0:01:22.679 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.066) 0:01:22.746 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.128) 0:01:22.875 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:139 Saturday 11 January 2025 11:29:57 -0500 (0:00:00.049) 0:01:22.924 ****** included: fedora.linux_system_roles.firewall for managed-node2 TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Saturday 11 January 2025 11:29:58 -0500 (0:00:00.134) 0:01:23.059 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed-node2 TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Saturday 11 January 2025 11:29:58 -0500 (0:00:00.069) 0:01:23.129 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Saturday 11 January 2025 11:29:58 -0500 (0:00:00.048) 0:01:23.177 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Saturday 11 January 2025 11:29:58 -0500 (0:00:00.041) 0:01:23.218 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22 Saturday 11 January 2025 11:29:58 -0500 (0:00:00.050) 0:01:23.269 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27 Saturday 11 January 2025 11:29:58 -0500 (0:00:00.063) 0:01:23.332 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 Saturday 11 January 2025 11:29:58 -0500 (0:00:00.045) 0:01:23.377 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: firewalld TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43 Saturday 11 January 2025 11:29:59 -0500 (0:00:00.767) 0:01:24.145 ****** skipping: [managed-node2] => { "false_condition": "__firewall_is_transactional | d(false)" } TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48 Saturday 11 January 2025 11:29:59 -0500 (0:00:00.127) 0:01:24.272 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53 Saturday 11 January 2025 11:29:59 -0500 (0:00:00.065) 0:01:24.338 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Saturday 11 January 2025 11:29:59 -0500 (0:00:00.064) 0:01:24.402 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Saturday 11 January 2025 11:29:59 -0500 (0:00:00.063) 0:01:24.466 ****** [WARNING]: TASK: fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services: The loop variable 'item' is already in use. You should set the `loop_var` value in the `loop_control` option for the task to something else to avoid variable collisions and unexpected behavior. skipping: [managed-node2] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Saturday 11 January 2025 11:29:59 -0500 (0:00:00.072) 0:01:24.539 ****** ok: [managed-node2] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sat 2025-01-11 11:29:10 EST", "ActiveEnterTimestampMonotonic": "271605433", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus-broker.service dbus.socket sysinit.target polkit.service system.slice basic.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sat 2025-01-11 11:29:10 EST", "AssertTimestampMonotonic": "271333792", "Before": "multi-user.target shutdown.target network-pre.target", "BindLogSockets": "no", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "418206000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanLiveMount": "yes", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2025-01-11 11:29:10 EST", "ConditionTimestampMonotonic": "271333789", "ConfigurationDirectoryMode": "0755", "Conflicts": "iptables.service ipset.service ip6tables.service ebtables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4516", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DebugInvocation": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698229248", "EffectiveMemoryMax": "3698229248", "EffectiveTasksMax": "22365", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainHandoffTimestampMonotonic": "271363582", "ExecMainPID": "10862", "ExecMainStartTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainStartTimestampMonotonic": "271336631", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2025-01-11 11:29:10 EST", "InactiveExitTimestampMonotonic": "271337679", "InvocationID": "f2f268445be048bc90d6f558b31f85a8", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13978", "LimitNPROCSoft": "13978", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13978", "LimitSIGPENDINGSoft": "13978", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LiveMountResult": "success", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "10862", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureDurationUSec": "[not set]", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3240796160", "MemoryCurrent": "32980992", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "35106816", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "0", "MemorySwapMax": "infinity", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivatePIDs": "no", "PrivateTmp": "no", "PrivateTmpEx": "no", "PrivateUsers": "no", "PrivateUsersEx": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectControlGroupsEx": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket sysinit.target system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestamp": "Sat 2025-01-11 11:29:10 EST", "StateChangeTimestampMonotonic": "271605433", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22365", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Saturday 11 January 2025 11:30:00 -0500 (0:00:00.595) 0:01:25.134 ****** ok: [managed-node2] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sat 2025-01-11 11:29:10 EST", "ActiveEnterTimestampMonotonic": "271605433", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus-broker.service dbus.socket sysinit.target polkit.service system.slice basic.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sat 2025-01-11 11:29:10 EST", "AssertTimestampMonotonic": "271333792", "Before": "multi-user.target shutdown.target network-pre.target", "BindLogSockets": "no", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "418206000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanLiveMount": "yes", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2025-01-11 11:29:10 EST", "ConditionTimestampMonotonic": "271333789", "ConfigurationDirectoryMode": "0755", "Conflicts": "iptables.service ipset.service ip6tables.service ebtables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4516", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DebugInvocation": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698229248", "EffectiveMemoryMax": "3698229248", "EffectiveTasksMax": "22365", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainHandoffTimestampMonotonic": "271363582", "ExecMainPID": "10862", "ExecMainStartTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainStartTimestampMonotonic": "271336631", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2025-01-11 11:29:10 EST", "InactiveExitTimestampMonotonic": "271337679", "InvocationID": "f2f268445be048bc90d6f558b31f85a8", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13978", "LimitNPROCSoft": "13978", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13978", "LimitSIGPENDINGSoft": "13978", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LiveMountResult": "success", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "10862", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureDurationUSec": "[not set]", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3240222720", "MemoryCurrent": "32980992", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "35106816", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "0", "MemorySwapMax": "infinity", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivatePIDs": "no", "PrivateTmp": "no", "PrivateTmpEx": "no", "PrivateUsers": "no", "PrivateUsersEx": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectControlGroupsEx": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket sysinit.target system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestamp": "Sat 2025-01-11 11:29:10 EST", "StateChangeTimestampMonotonic": "271605433", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22365", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Saturday 11 January 2025 11:30:00 -0500 (0:00:00.561) 0:01:25.696 ****** ok: [managed-node2] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3.12", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Saturday 11 January 2025 11:30:00 -0500 (0:00:00.061) 0:01:25.758 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Saturday 11 January 2025 11:30:00 -0500 (0:00:00.043) 0:01:25.802 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Saturday 11 January 2025 11:30:00 -0500 (0:00:00.044) 0:01:25.846 ****** [WARNING]: TASK: fedora.linux_system_roles.firewall : Configure firewall: The loop variable 'item' is already in use. You should set the `loop_var` value in the `loop_control` option for the task to something else to avoid variable collisions and unexpected behavior. ok: [managed-node2] => (item={'port': '15001-15003/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "15001-15003/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Saturday 11 January 2025 11:30:01 -0500 (0:00:00.515) 0:01:26.361 ****** [WARNING]: TASK: fedora.linux_system_roles.firewall : Gather firewall config information: The loop variable 'item' is already in use. You should set the `loop_var` value in the `loop_control` option for the task to something else to avoid variable collisions and unexpected behavior. skipping: [managed-node2] => (item={'port': '15001-15003/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "'detailed' in fw[0]", "item": { "port": "15001-15003/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed-node2] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Saturday 11 January 2025 11:30:01 -0500 (0:00:00.058) 0:01:26.420 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'detailed' in fw[0]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Saturday 11 January 2025 11:30:01 -0500 (0:00:00.101) 0:01:26.522 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Saturday 11 January 2025 11:30:01 -0500 (0:00:00.041) 0:01:26.564 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Saturday 11 January 2025 11:30:01 -0500 (0:00:00.053) 0:01:26.618 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Saturday 11 January 2025 11:30:01 -0500 (0:00:00.086) 0:01:26.705 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Saturday 11 January 2025 11:30:01 -0500 (0:00:00.086) 0:01:26.791 ****** skipping: [managed-node2] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:146 Saturday 11 January 2025 11:30:01 -0500 (0:00:00.082) 0:01:26.873 ****** redirecting (type: modules) ansible.builtin.selinux to ansible.posix.selinux redirecting (type: modules) ansible.builtin.selinux to ansible.posix.selinux redirecting (type: modules) ansible.builtin.seboolean to ansible.posix.seboolean included: fedora.linux_system_roles.selinux for managed-node2 TASK [fedora.linux_system_roles.selinux : Set ansible_facts required by role and install packages] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:2 Saturday 11 January 2025 11:30:02 -0500 (0:00:00.154) 0:01:27.028 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml for managed-node2 TASK [fedora.linux_system_roles.selinux : Ensure ansible_facts used by role] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml:2 Saturday 11 January 2025 11:30:02 -0500 (0:00:00.068) 0:01:27.097 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Ensure SELinux packages] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml:7 Saturday 11 January 2025 11:30:02 -0500 (0:00:00.128) 0:01:27.225 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml for managed-node2 TASK [fedora.linux_system_roles.selinux : Check if system is ostree] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:5 Saturday 11 January 2025 11:30:02 -0500 (0:00:00.123) 0:01:27.349 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:10 Saturday 11 January 2025 11:30:02 -0500 (0:00:00.063) 0:01:27.413 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:17 Saturday 11 January 2025 11:30:02 -0500 (0:00:00.055) 0:01:27.468 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:22 Saturday 11 January 2025 11:30:02 -0500 (0:00:00.052) 0:01:27.521 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux python2 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:26 Saturday 11 January 2025 11:30:02 -0500 (0:00:00.049) 0:01:27.570 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_python_version is version('3', '<')", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux python3 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:35 Saturday 11 January 2025 11:30:02 -0500 (0:00:00.042) 0:01:27.613 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: python3-libselinux python3-policycoreutils TASK [fedora.linux_system_roles.selinux : Install SELinux python3 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:46 Saturday 11 January 2025 11:30:03 -0500 (0:00:00.784) 0:01:28.397 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_os_family == \"Suse\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux tool semanage] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:58 Saturday 11 January 2025 11:30:03 -0500 (0:00:00.053) 0:01:28.451 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: policycoreutils-python-utils TASK [fedora.linux_system_roles.selinux : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:72 Saturday 11 January 2025 11:30:04 -0500 (0:00:00.792) 0:01:29.243 ****** skipping: [managed-node2] => { "false_condition": "__selinux_is_transactional | d(false)" } TASK [fedora.linux_system_roles.selinux : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:77 Saturday 11 January 2025 11:30:04 -0500 (0:00:00.064) 0:01:29.308 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:82 Saturday 11 January 2025 11:30:04 -0500 (0:00:00.102) 0:01:29.411 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Refresh facts] *********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:89 Saturday 11 January 2025 11:30:04 -0500 (0:00:00.049) 0:01:29.460 ****** ok: [managed-node2] TASK [fedora.linux_system_roles.selinux : Set permanent SELinux state if enabled] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:5 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.887) 0:01:30.347 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_selinux.status == \"enabled\" and (selinux_state or selinux_policy)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set permanent SELinux state if disabled] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:13 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.047) 0:01:30.395 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_selinux.status == \"disabled\" and selinux_state", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set selinux_reboot_required] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:21 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.043) 0:01:30.438 ****** ok: [managed-node2] => { "ansible_facts": { "selinux_reboot_required": false }, "changed": false } TASK [fedora.linux_system_roles.selinux : Fail if reboot is required] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:25 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.053) 0:01:30.492 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_reboot_required", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Warn if SELinux is disabled] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:30 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.039) 0:01:30.531 ****** skipping: [managed-node2] => { "false_condition": "ansible_selinux.status == \"disabled\"" } TASK [fedora.linux_system_roles.selinux : Drop all local modifications] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:35 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.036) 0:01:30.567 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_all_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux boolean local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:43 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.035) 0:01:30.603 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_booleans_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux file context local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:48 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.039) 0:01:30.643 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_fcontexts_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux port local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:53 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.038) 0:01:30.682 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_ports_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux login local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:58 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.041) 0:01:30.723 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_logins_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set SELinux booleans] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:63 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.094) 0:01:30.818 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Set SELinux file contexts] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:74 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.042) 0:01:30.860 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Set an SELinux label on a port] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:87 Saturday 11 January 2025 11:30:05 -0500 (0:00:00.041) 0:01:30.902 ****** ok: [managed-node2] => (item={'ports': '15001-15003', 'setype': 'http_port_t'}) => { "__selinux_item": { "ports": "15001-15003", "setype": "http_port_t" }, "ansible_loop_var": "__selinux_item", "changed": false, "ports": [ "15001-15003" ], "proto": "tcp", "setype": "http_port_t", "state": "present" } TASK [fedora.linux_system_roles.selinux : Set linux user to SELinux user mapping] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:99 Saturday 11 January 2025 11:30:06 -0500 (0:00:00.622) 0:01:31.525 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Get SELinux modules facts] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:112 Saturday 11 January 2025 11:30:06 -0500 (0:00:00.040) 0:01:31.565 ****** ok: [managed-node2] => { "ansible_facts": { "selinux_checksums": true, "selinux_installed_modules": { "abrt": { "100": { "checksum": "sha256:5fdd1acb3c17ce67f51edc9160653adada9aade378218aed1b38acc0b4d4157a", "enabled": 1 } }, "accountsd": { "100": { "checksum": "sha256:50197ed40bbac67de44c12d745bc4a1259c8ae3f3828f5b4cfcd02b187ee1f47", "enabled": 1 } }, "acct": { "100": { "checksum": "sha256:14149edb1593dad6e684a672c57ca35f2ff51aee24ef503171c28671dc1bff13", "enabled": 1 } }, "afs": { "100": { "checksum": "sha256:769aaf25a8699233ceeda87ac858ebc91f094ea98b303c54e2c810760acd9756", "enabled": 1 } }, "afterburn": { "100": { "checksum": "sha256:9b9c00f12dea94989d2f5c592fa8ba4cff49b1ae7c9636fc38baa7957750e80f", "enabled": 1 } }, "aiccu": { "100": { "checksum": "sha256:d7ed1435298abcb42678af33c66ba2bac4c02723601c5018ecef64ca4c4e8f0b", "enabled": 1 } }, "aide": { "100": { "checksum": "sha256:95c6046b0d690df4792fe4418de5f7753279b72c78c5f1dab5397aba8bdf7152", "enabled": 1 } }, "alsa": { "100": { "checksum": "sha256:c0be471cd7182d1126cd6d19b276742a6ace62429425a81d42d8af4c24bcdbb4", "enabled": 1 } }, "amanda": { "100": { "checksum": "sha256:2ea0f86264cad0fa7f1eba0e6238526e92fadc65fd9754f62bc6d59b2da9086c", "enabled": 1 } }, "amtu": { "100": { "checksum": "sha256:1c77cc13007a0024e3121f064aad7dafce5fd57d7f71140f0af6b5af10805a8b", "enabled": 1 } }, "anaconda": { "100": { "checksum": "sha256:79386dbc8f95842f819ddbb01ce5824752d78bb4992783288d79db391a4ec3b5", "enabled": 1 } }, "antivirus": { "100": { "checksum": "sha256:48aafc3e308b4db80797ec75f52d3cc1b8216315d28108a10a03e47a0209adda", "enabled": 1 } }, "apache": { "100": { "checksum": "sha256:30a1e157e56f57a82557524173635e911f796fa8733f98ad190fc854aba893d8", "enabled": 1 } }, "apcupsd": { "100": { "checksum": "sha256:46a2e84d4dd0064f0b368787ed9c494ace4e337d5d80bcc6db77f66dede34375", "enabled": 1 } }, "apm": { "100": { "checksum": "sha256:4f5ff9b6ad61bc6a311bc71768e7ead7fe273989b7cf2fa58b1a13221e8d82f4", "enabled": 1 } }, "application": { "100": { "checksum": "sha256:ee11bb52bc37177920d348c4b920738a1b64413f516c66e18fa5a453a6d81f5e", "enabled": 1 } }, "arpwatch": { "100": { "checksum": "sha256:11a3c765598ae81cda7af561ada4795fb53a9ae9cfcb02af806b08cfc7f00b0e", "enabled": 1 } }, "asterisk": { "100": { "checksum": "sha256:ce887eb790f85810eb5eec0ac96a4e8028d633cf16fc117db4f733e6040c11c2", "enabled": 1 } }, "auditadm": { "100": { "checksum": "sha256:201655fdd63b544aa448c7b46c8b7ec8aa726ef3b55e9ef928ff3502c128973b", "enabled": 1 } }, "authlogin": { "100": { "checksum": "sha256:f1937fa4d112f634f69b9f0104872ce507c56928951a961c26ddf385a12be54e", "enabled": 1 } }, "automount": { "100": { "checksum": "sha256:5d308d3654fc5b0cea77156c27b98bbe2e370b843a920d75e1d616472caee55f", "enabled": 1 } }, "avahi": { "100": { "checksum": "sha256:ab1c62da7d38c1870efff54480b9a43ac2683eb0ebd77925a6fdea0fd47e207c", "enabled": 1 } }, "awstats": { "100": { "checksum": "sha256:f4a83008d582e61321a1d5462807c402d6cd3034d272dab00dd788122f245114", "enabled": 1 } }, "bacula": { "100": { "checksum": "sha256:04e3c063e146bd0f96587e9f49cdb4020e83bf69ba77250d61d2c9fcea53476b", "enabled": 1 } }, "base": { "100": { "checksum": "sha256:106220779e72010bf08e45ebec252a2e47257ca1134c07dbcd68a0d37e790931", "enabled": 1 } }, "bcfg2": { "100": { "checksum": "sha256:860a81a71f68a66e7dc0cee62e02f5da8d2dbb83724bfde9abc24b56b06ec914", "enabled": 1 } }, "bind": { "100": { "checksum": "sha256:ab2767aacd75df866fc6080d7db8dfefbd013a4ad578412ccfa6dece0d0b8ea7", "enabled": 1 } }, "bitlbee": { "100": { "checksum": "sha256:6ae2a654612810131bc874411156b1ef5d784143484e2a867ef413b6ab0f8750", "enabled": 1 } }, "blkmapd": { "100": { "checksum": "sha256:a6ac037c876fcea13d9457717138ec1c17913a0a303469219aa610351225c30c", "enabled": 1 } }, "blueman": { "100": { "checksum": "sha256:10372514dd684076a8cc4c93211d6f6732dcff0fa6aceed02dce868b12aaee82", "enabled": 1 } }, "bluetooth": { "100": { "checksum": "sha256:d8e9f771f666c02cacd7881f83cebd64ca88d79382fddca9e608d75df65099a2", "enabled": 1 } }, "boinc": { "100": { "checksum": "sha256:4ede0a1fb4501325ad8d0ff576f2f20bb0c3e994d2142372449de840f10e8b4b", "enabled": 1 } }, "boltd": { "100": { "checksum": "sha256:63a74e88f61411cfd03aa6e719867d1ae3238b99f3adb2f91cc0973e8b10e174", "enabled": 1 } }, "boothd": { "100": { "checksum": "sha256:4df95e4f92b1da81af245d74fd4481aad048554cad1d31f87985e3c68814ac8e", "enabled": 1 } }, "bootloader": { "100": { "checksum": "sha256:c4b351c2ad386a4c22a477788aeba9ae713cddaae623d733f1957d5d521c717d", "enabled": 1 } }, "bootupd": { "100": { "checksum": "sha256:fd688c68b40c042dd04fe9cb65c7a270353a2f7c677a5e73864b34ee6118f94d", "enabled": 1 } }, "brctl": { "100": { "checksum": "sha256:3a931c1c282cee1b53f6aee3730fef0e498fba3c5005218b4c7a923a7ce226c2", "enabled": 1 } }, "brltty": { "100": { "checksum": "sha256:0471f4ad4a09025a1a23ad398d869d88b90172a869800cf06660c886033c7409", "enabled": 1 } }, "bugzilla": { "100": { "checksum": "sha256:480fe3690a301d2359b747a03b7f65fd881063f6f7d877c55a16e6a51f324df6", "enabled": 1 } }, "cachefilesd": { "100": { "checksum": "sha256:04eafc0b0c7783758bba1cb55893416c204933bf0fdef964ce23c6f17c626c65", "enabled": 1 } }, "calamaris": { "100": { "checksum": "sha256:3b369138db2945167e6b10503d99902411fd5213b91dae978afe344b174240f7", "enabled": 1 } }, "callweaver": { "100": { "checksum": "sha256:0d66c1686efea20f479f6dd3057f5bd5550709d34e3a30745900abf363728bf1", "enabled": 1 } }, "canna": { "100": { "checksum": "sha256:5677c14448ec90dd564b7fe2c19146faa264a6cc80da75dba8f51c1e17af73d2", "enabled": 1 } }, "ccs": { "100": { "checksum": "sha256:ab0abd34c688230d836833a4f46228f3873ccbcfd3fd351d1306c203368ab63b", "enabled": 1 } }, "cdrecord": { "100": { "checksum": "sha256:c258276b231b731865f4237c8477010ea00b71e0e51501d7a206344381f625b1", "enabled": 1 } }, "certmaster": { "100": { "checksum": "sha256:5d46915b4fa32deccdf587b5970cde875d7b43643c09ec42a156193f9592c8ab", "enabled": 1 } }, "certmonger": { "100": { "checksum": "sha256:8746abfbb345c4bfd16b1c27e127a281e72388536aa9c7545ed308a266aceeb5", "enabled": 1 } }, "certwatch": { "100": { "checksum": "sha256:8d8299ad27c884a1657ff8ba52413f24867605e5e31c55a70659b3b0db4309d5", "enabled": 1 } }, "cfengine": { "100": { "checksum": "sha256:c3e0da50f4e8cd55d3a139afdcfa0fc5479b0e4afe78df4a1eab9400a71b64e8", "enabled": 1 } }, "cgroup": { "100": { "checksum": "sha256:e681dc6218ef1387435a6c4e95ed13e45d5edf2089991beb9bc8df41021cd865", "enabled": 1 } }, "chrome": { "100": { "checksum": "sha256:d64d5c77ef6d19be3aecf839fc444fab44a5717fa879c6d1bf6eb7564f556835", "enabled": 1 } }, "chronyd": { "100": { "checksum": "sha256:ea2b23dec628aecd68d6b8bc7d88d662e6bc0e72330f2a87e436f66a38b28ed4", "enabled": 1 } }, "cifsutils": { "100": { "checksum": "sha256:db66a0e17ecf7229eff4ffe84885bbff334085e3e20c282dd7e1eab97545e5e7", "enabled": 1 } }, "cinder": { "100": { "checksum": "sha256:95351f4464622aed05297db68b883d47165ce7db322134048f5a99479c734da3", "enabled": 1 } }, "cipe": { "100": { "checksum": "sha256:25016148cbfa7480cee4b614572bf7a364f6e45a5aca844bb9a87bba217b93a5", "enabled": 1 } }, "clock": { "100": { "checksum": "sha256:5f8b31db8708926fe3de99aadf34b5114184bff20956c7203c6329433bc24dc7", "enabled": 1 } }, "clogd": { "100": { "checksum": "sha256:b6bdd15aeae4574bffcd8a3c862261eb9e21ae5451035cc470ee9eba6cf49cd4", "enabled": 1 } }, "cloudform": { "100": { "checksum": "sha256:8d891dca8241582619e85b4d6f5aa273bd87d3bb7ff04199a38c91f47df12403", "enabled": 1 } }, "cmirrord": { "100": { "checksum": "sha256:e9ccb95f73f750fa464aa4fb5b76e5bad42050e27a128795826272ace5f70760", "enabled": 1 } }, "cobbler": { "100": { "checksum": "sha256:e090b7686819373bcad8a0cdfcf17dd453f7873dfe757742b4bee5c2be7f58d5", "enabled": 1 } }, "collectd": { "100": { "checksum": "sha256:96c655a73312da543bc2aa652b14079cb6fdac677618325d9f1ad83996dbeee1", "enabled": 1 } }, "colord": { "100": { "checksum": "sha256:180a83a6a6059ef9044f9a171ec71a5934a5a19da0f9ce5451c3ba1bbfb75343", "enabled": 1 } }, "comsat": { "100": { "checksum": "sha256:d1e10fcd329ec130a7346374dd6cf339eb38e2c37bc05f371c0381cc7e022d2e", "enabled": 1 } }, "condor": { "100": { "checksum": "sha256:f27781ec40ba6625c3854cf0e6c311a345d5619a3b07c7105635dc60a9369a6c", "enabled": 1 } }, "conman": { "100": { "checksum": "sha256:7984cbaf5572913966bae94db45941e7e15fc195f70c0bd30cbd6b8a24998687", "enabled": 1 } }, "conntrackd": { "100": { "checksum": "sha256:0607c19ac5763806c4420869caf11320ecad25b9ba9e10307105c75e9a26f6be", "enabled": 1 } }, "consolekit": { "100": { "checksum": "sha256:e6e39c814a93fa8973beffb8f891030f7b8094d9987e221d140c1a1fc79a196b", "enabled": 1 } }, "coreos_installer": { "100": { "checksum": "sha256:0ca9390a1bd0c5caa0958c03c1f665a129836f50e3ee11d54fa21800fab9aa08", "enabled": 1 } }, "couchdb": { "100": { "checksum": "sha256:e7ffb1afd19042410ffc4af01ee0fe31ff9f434d23451dd3bd01a7052be716b2", "enabled": 1 } }, "courier": { "100": { "checksum": "sha256:113abb8c428ab21653d9c2f9aef35354470381234ccb493e882ef22f72c2ede4", "enabled": 1 } }, "cpucontrol": { "100": { "checksum": "sha256:36939ecc2515a7980f3be7a2d9518018b3502aa526a446b04552f43a671518dc", "enabled": 1 } }, "cpufreqselector": { "100": { "checksum": "sha256:43b7f9d92ade2e28e26c7f9736d94d91b9a20812118f466b39aa4dc7a02e0959", "enabled": 1 } }, "cpuplug": { "100": { "checksum": "sha256:97a1c9de6df462cab0645b58e16072042c4cab7704c10209950684673a049b8d", "enabled": 1 } }, "cron": { "100": { "checksum": "sha256:0e02e661cc66e751e27d9ece16db633d5ab2036dab470b9dd4a1ef0725bf1534", "enabled": 1 } }, "ctdb": { "100": { "checksum": "sha256:7a7a3998f8a29178005386630556039bba9812f206b5f6782bdcfc0c22f66cbc", "enabled": 1 } }, "cups": { "100": { "checksum": "sha256:90605693f95fba47762dba603d99f0603ccea7238937bd0ed4ca8dc680fd7382", "enabled": 1 } }, "cvs": { "100": { "checksum": "sha256:05274fef587a5ac9def4bce8be2540bc893dd2c7b0052ad9f4e5de0c6da55718", "enabled": 1 } }, "cyphesis": { "100": { "checksum": "sha256:635739376db164b36a935a3bb0dbe8b5644b0fca8d3791b0efde3da36372a4aa", "enabled": 1 } }, "cyrus": { "100": { "checksum": "sha256:26567d47104e4634ab34aa39e0a3ea5f4ed85670b6c66b63e86c47c809080344", "enabled": 1 } }, "daemontools": { "100": { "checksum": "sha256:429c5e1b65232dbfb74741ba1514315288e2704c8a105759c31b85e3742200e1", "enabled": 1 } }, "dbadm": { "100": { "checksum": "sha256:24d6e3dfb505a297328108979f63cbad1b2a0f241c2fb2e84f93dadc086f1273", "enabled": 1 } }, "dbskk": { "100": { "checksum": "sha256:0bf031897570904400ae864c9d12f8880047d1c0368ebb8edc7fd7232f2eecab", "enabled": 1 } }, "dbus": { "100": { "checksum": "sha256:c0b2fa3d02d462ddbf8269485861a3e6e873cd5f05a174a99b121095cbd31f97", "enabled": 1 } }, "dcc": { "100": { "checksum": "sha256:f237213b8df023a69261f26318e3572f28a2c03bccf8d37505dabeca64aa9fe4", "enabled": 1 } }, "ddclient": { "100": { "checksum": "sha256:4159c1afa1a0e1977328780489eb61c30f61823b389f4cd0d79f25a7b27d5857", "enabled": 1 } }, "denyhosts": { "100": { "checksum": "sha256:643ce1a1f7dc2df66e3a1ff57d3dbd9631ea7c9b9b83467d7e4e2f4dc701186b", "enabled": 1 } }, "devicekit": { "100": { "checksum": "sha256:845cf9ac75f705a72df5d2b643b810fe266919369f6840e461f9d3043db97bb9", "enabled": 1 } }, "dhcp": { "100": { "checksum": "sha256:57fe91a9ccded7037429132ca5c28c408881f296eb8e603041b61f82c673c7c6", "enabled": 1 } }, "dictd": { "100": { "checksum": "sha256:246922587bf734f34d26ddb793ab9c649460b3cea63acfca54dc8a989e948d97", "enabled": 1 } }, "dirsrv": { "100": { "checksum": "sha256:9c67bc04c2d69254c0421c864b6f309c4c213bf415759b93a11a21bafa55b7f2", "enabled": 1 } }, "distcc": { "100": { "checksum": "sha256:a8179c2e64684b994ca8802badba7dbe0df42564f400307eeca0274f442cc0c4", "enabled": 1 } }, "dmesg": { "100": { "checksum": "sha256:977de79bba759fc421b791cc5496b0479b34077259eb0b9aab7049cce9e3d537", "enabled": 1 } }, "dmidecode": { "100": { "checksum": "sha256:dbe77a5d3671052e5b1dff63016c24e6e576803c82edbf2a913147a24f893124", "enabled": 1 } }, "dnsmasq": { "100": { "checksum": "sha256:7ad2d3372989319aaef08ac68aff266ae10dd262728284d4531851e893fe9edf", "enabled": 1 } }, "dnssec": { "100": { "checksum": "sha256:70ff660118b3bb9c8e5107d566adf6c2ad0b4af94d6757852c68a00eb19f6e56", "enabled": 1 } }, "dovecot": { "100": { "checksum": "sha256:1057c6ca706e9266c2e147865fb58e634439b7ec3e6c1faba0185401e2d4e985", "enabled": 1 } }, "drbd": { "100": { "checksum": "sha256:225633f600a5c31fcb19a9b721677833e32b88e57ef8cfaa3fa88b25bfc87c87", "enabled": 1 } }, "dspam": { "100": { "checksum": "sha256:503e9b509a0f764faea4a2415311aa1b43bfd8b78bbcc8090f6bc81c200beea8", "enabled": 1 } }, "entropyd": { "100": { "checksum": "sha256:e2aa3a7bc12f543b79842c2504cb1b53b73c55c78a42e24fbead6ab2deb120a4", "enabled": 1 } }, "exim": { "100": { "checksum": "sha256:b097146234c292a604496276a05ae2cbb09bbfdbf0ac9d7948b42afddd953347", "enabled": 1 } }, "fail2ban": { "100": { "checksum": "sha256:c51270c9abffd1d173532bbce08b28480f8e261d099370268b5364c343726140", "enabled": 1 } }, "fcoe": { "100": { "checksum": "sha256:a8a2c86fcc37532b9bf516e070ff53c9e675cb5e9f154966dde6a9b5fc6d7bc5", "enabled": 1 } }, "fdo": { "100": { "checksum": "sha256:7ecdacdeaddd502ea66686e0eaa24fa19fef7202f0092a7174db2aa86eddae8e", "enabled": 1 } }, "fedoratp": { "100": { "checksum": "sha256:85ee3ad2c50bf2b0ba1fb4fe0c57850f9ffd3bb5c11d11a5b5e159ac65783c18", "enabled": 1 } }, "fetchmail": { "100": { "checksum": "sha256:56ef87427ba73e1981192f595dd2c169b322cbed91da8de818e995434e07c757", "enabled": 1 } }, "finger": { "100": { "checksum": "sha256:9aa41b846e607a383c12d0bb4a9695058f8c9b1ed4b26bb59a96855015d9ab35", "enabled": 1 } }, "firewalld": { "100": { "checksum": "sha256:9e85ce564ebec8cae461d093727c4e339a32b10ce3efa569b5c2d53c27baae59", "enabled": 1 } }, "firewallgui": { "100": { "checksum": "sha256:433e16150ab0650ca51df171c1fbe1406cfe15dfa077bade2e48a7d8514bb0ba", "enabled": 1 } }, "firstboot": { "100": { "checksum": "sha256:a16615a4aea8b12fce51f6d7715627da3c3f6c6099db8c3087e29c920d81f861", "enabled": 1 } }, "fprintd": { "100": { "checksum": "sha256:1cb012beb208aebe26b53919c54e8e902210feb7b26d52b05cbe760d273b0645", "enabled": 1 } }, "freeipmi": { "100": { "checksum": "sha256:f62cf10a14cb8de0f158fed8cc9400a8769be1ffb2f3dcdcd0e920de20d68007", "enabled": 1 } }, "freqset": { "100": { "checksum": "sha256:5c96b7146bde0acec28d73afc25b0a35fbfffe2ccd69529717eca6927dc558a7", "enabled": 1 } }, "fstools": { "100": { "checksum": "sha256:0fdc4babffbc8fa18d9445b74b14fc08b77fd6e371c96dc1f5283d6ff7c5f546", "enabled": 1 } }, "ftp": { "100": { "checksum": "sha256:dae3209382de68fdb49a72792108ec3a55c0e50abdfd38c5fae300da4c561c47", "enabled": 1 } }, "fwupd": { "100": { "checksum": "sha256:4f1c84d8ccd64b0af49edc9247ac92f49545fe4dfb24906acc133ebd7ba78bb9", "enabled": 1 } }, "games": { "100": { "checksum": "sha256:7d9b09309f694125d4ba8e1f5683baedfc2c935f0242be94531b43d842509ae9", "enabled": 1 } }, "gdomap": { "100": { "checksum": "sha256:bec0b003cf82cc2fe7e6f196bd82b929e87a827fe8d46bca5b1df2613a409d6b", "enabled": 1 } }, "geoclue": { "100": { "checksum": "sha256:3e8a90189684079f66e782c8e001c435ada771766dd1fa9ae33a977405f6ba99", "enabled": 1 } }, "getty": { "100": { "checksum": "sha256:e5ecf805b71dfe0f039f8eeefa730893c5417fe53b0e0eeba55f99670b4c4584", "enabled": 1 } }, "git": { "100": { "checksum": "sha256:69850d2342f0fcb96a719d39a3698d547c105880e6834f232a4d845a18a046f3", "enabled": 1 } }, "gitosis": { "100": { "checksum": "sha256:24935d3495375c7f6e0a0ddc091cf1f3f276ee155d07760bd7ba0c3ac115f6e8", "enabled": 1 } }, "glance": { "100": { "checksum": "sha256:fe00543de47f940ba5bf1ab84b7dc7f6a680053ef148329f2a9ec827eb47c61c", "enabled": 1 } }, "glusterd": { "100": { "checksum": "sha256:4d2befdcd924820622a0c6653772a9f5ba96aad9a3bd7b4d86daab96b233c9ea", "enabled": 1 } }, "gnome": { "100": { "checksum": "sha256:a5e2adeeda1a5b359025d956cbf20eb3773341207c72cdf7bf5001deeedf4712", "enabled": 1 } }, "gnome_remote_desktop": { "100": { "checksum": "sha256:d926eb26b34a6253b249bc1cea7696dae87887739ef77cd8cb92bdf7f6d1f371", "enabled": 1 } }, "gpg": { "100": { "checksum": "sha256:a9180cca4098d53514cd5f701f4afd40420a3664e8ffa86797ee3a8fadd9a84a", "enabled": 1 } }, "gpm": { "100": { "checksum": "sha256:2b5e83f7513a2470766cbaad46e28604ed8fcd6b640dc0350d90731a807f56ad", "enabled": 1 } }, "gpsd": { "100": { "checksum": "sha256:b64db8adeb9e8602211a0924c1b4b917d0f701df86984e73412344c9b3c2f5d9", "enabled": 1 } }, "gssproxy": { "100": { "checksum": "sha256:a43f376f2b1605cf0d45884b818db045eb37b4f037ae7af36ec5bf993676f4e5", "enabled": 1 } }, "guest": { "100": { "checksum": "sha256:262ccc7c3b237e7ebc31631d83baf2c6cbf8438cce02ce82f008fd6a3ff81a43", "enabled": 1 } }, "hddtemp": { "100": { "checksum": "sha256:fc86a156d59e31b02ce40cbda3c24b0c0baae2ed6f0cabf84cb9a3248ccd236e", "enabled": 1 } }, "hostapd": { "100": { "checksum": "sha256:ba13f04311a741f26ac34cd39688e2f4e75a546358c63c57a2d6a916dfc18315", "enabled": 1 } }, "hostname": { "100": { "checksum": "sha256:7366a48a1774b579953d8b8ae05f9ebe07e5d27984c995afe27509eacf1d79d6", "enabled": 1 } }, "hsqldb": { "100": { "checksum": "sha256:a0c31e8462c7fbbc6579270e5e35923deff56790f565157cb9c32c4eb3d6958b", "enabled": 1 } }, "hwloc": { "100": { "checksum": "sha256:fde041b834e5939945fc0d91e45940745254e128d96625ca5ab4d6e7a4379e27", "enabled": 1 } }, "hypervkvp": { "100": { "checksum": "sha256:520caa8d38f07a36d096f7ee8eb1b0610f49eafedca1721b340f5cdf21e23944", "enabled": 1 } }, "ibacm": { "100": { "checksum": "sha256:4bc0f9a40691f1abb15ce6cd277d7bab4b18720373a197c8310aa63995bcda0f", "enabled": 1 } }, "ica": { "100": { "checksum": "sha256:5daeb045f8b76fed05c1145d131f77dd556de131f777b7e259d8ec57787e7da6", "enabled": 1 } }, "icecast": { "100": { "checksum": "sha256:281ed7601c5b057802da5ffdc24a7e3308e0bb2591a263186f51e35816dffc12", "enabled": 1 } }, "iiosensorproxy": { "100": { "checksum": "sha256:05dd0967ab5bc2422ac57f2ec8a354181bf7038f0d1aabc2775c7c915bbbeaf3", "enabled": 1 } }, "inetd": { "100": { "checksum": "sha256:52f6df31e38b0c20b9fc264115ce56b4d6d8048dce5b6f250a6ca1b0675a0569", "enabled": 1 } }, "init": { "100": { "checksum": "sha256:7141bbe724e0c7f4b3dd7f335ab21267676b126f0443da9cbfa58f1372deb59c", "enabled": 1 } }, "inn": { "100": { "checksum": "sha256:78ae9628e756da216964c88e987dd91e40ec2803ff209763f628a434c8bf3b90", "enabled": 1 } }, "insights_client": { "100": { "checksum": "sha256:c6e547e12ca21f1f68ba74d667fab1bcbfc8d0461fc3a560e80160f32638ddba", "enabled": 1 } }, "iodine": { "100": { "checksum": "sha256:5c8ef0d8e730a3baca51f4c64b13df56149ef304effdc71fe0fb63f90e5a7296", "enabled": 1 } }, "iotop": { "100": { "checksum": "sha256:114de54fd32be9c32e88348ddb168291c0aaaba3a7937501f49e514585228092", "enabled": 1 } }, "ipmievd": { "100": { "checksum": "sha256:5777d2cc5e62942022a8e914847dc7436264585e41e2ce703c49c735e44add8c", "enabled": 1 } }, "ipsec": { "100": { "checksum": "sha256:ffde25d3dd0fa80b933b30cd457ea900389f84d13d4ea629f54c0be59136d4a6", "enabled": 1 } }, "iptables": { "100": { "checksum": "sha256:4c92a7102c209b0edf571a3ca3b35c597bd427d0d258c6829bfd29358448598b", "enabled": 1 } }, "irc": { "100": { "checksum": "sha256:291a822392f8a93d820aa03613bd0c058cfd1118fd6cca93a12175456eb2cf72", "enabled": 1 } }, "irqbalance": { "100": { "checksum": "sha256:2fc17d69e2cba360a0d9043585a3799f86697bf9189fcde62346b41bcf5bd9d9", "enabled": 1 } }, "iscsi": { "100": { "checksum": "sha256:a91d9d18f0502679892047e7a6023107b77412ae4b3d25f75982f766c792be58", "enabled": 1 } }, "isns": { "100": { "checksum": "sha256:5372fd0c432735b6ec28e0a5cafeebd74688912c7703616ca46b77287385c564", "enabled": 1 } }, "jabber": { "100": { "checksum": "sha256:d1916b00ece1c2b546cc04a7b6695f6459724a6bb7a92be78f399f61a613af86", "enabled": 1 } }, "jetty": { "100": { "checksum": "sha256:c78ba53bd379a65017e29895da717fa8e758494e94113b7f324d73121cb35423", "enabled": 1 } }, "jockey": { "100": { "checksum": "sha256:4cee27b59fe57c672587d220f3061417207a02bc0ff36188148f5c5f3101c1ad", "enabled": 1 } }, "journalctl": { "100": { "checksum": "sha256:27aeef17f22b9e891c4c793870f1124ec69ec7dc8f532bfec5b2a0e7319866cb", "enabled": 1 } }, "kafs": { "100": { "checksum": "sha256:602c3c40b28b653e11110e83c7f63122a75b9ad5cb7605299f6951a485284727", "enabled": 1 } }, "kdump": { "100": { "checksum": "sha256:bac33f3ca081f5f4b58fbbdbd3630aef7fb77c19368ec2cef66494cb8ef7fd90", "enabled": 1 } }, "kdumpgui": { "100": { "checksum": "sha256:7e16a86780b3b3d8a6afc5a09335cd6bf6bb39eec53815af9bbc53844655846c", "enabled": 1 } }, "keepalived": { "100": { "checksum": "sha256:5e76b1d8c27eea054956c86458f44ebc46d803035e2b3c0af67fd5feae26b9f2", "enabled": 1 } }, "kerberos": { "100": { "checksum": "sha256:65a81f3d94068c6e7cd3c1092cfbc080f853c7e02490356f7f16d40a3e8ace49", "enabled": 1 } }, "keyboardd": { "100": { "checksum": "sha256:e267709b6fc5f39679c3156b764ac82facb0b228e5b300d891f4a182662a1a5b", "enabled": 1 } }, "keystone": { "100": { "checksum": "sha256:1a7021842ad7350bc5674a455f4232b831ec8153d14635eba8930e8ade0980b7", "enabled": 1 } }, "keyutils": { "100": { "checksum": "sha256:0dc521ba68e3d21396cf5bce20427e6c0c5dfe635c743ccd4374beadd1a96cc5", "enabled": 1 } }, "kismet": { "100": { "checksum": "sha256:4398910d875745fbd61926116eddc0c7f3f6da3cf8b62fb4149d78734c344df3", "enabled": 1 } }, "kpatch": { "100": { "checksum": "sha256:36ae9280a2f7e86befeaf301f1d47087b9214bc97e79562105fcaa1246512467", "enabled": 1 } }, "ksmtuned": { "100": { "checksum": "sha256:66f6fee140f5f767906ea257b32ede5d214b8da0db4031313eb1fddb9a589d21", "enabled": 1 } }, "ktalk": { "100": { "checksum": "sha256:e783786109e8dac25e8be80e5782ec260343cf0db5b03126b85811a101bd9d0e", "enabled": 1 } }, "l2tp": { "100": { "checksum": "sha256:55f189e95389ac2f9205270f3dd53e25761d8e626989bdfb3c05b19c0086dc30", "enabled": 1 } }, "ldap": { "100": { "checksum": "sha256:cb97aa280c3bb13b6c8804e2a2bf971250db72b77ba5ece99702ab6434c3d14a", "enabled": 1 } }, "libraries": { "100": { "checksum": "sha256:bb45e05979c90858a0ed2ed0ecae78cc4e99b750145947a2940eb0493249d161", "enabled": 1 } }, "likewise": { "100": { "checksum": "sha256:f535c1e16dc83389fd5a2fa4716615c687a931e3a9ad7f65dc8607ef305f7994", "enabled": 1 } }, "linuxptp": { "100": { "checksum": "sha256:0d6f39eec4849cc1f5ca0014c29399b78317216ed0e73dfc524be3194ff1bc08", "enabled": 1 } }, "lircd": { "100": { "checksum": "sha256:cdd2ad3d9fa15b37c76e5bf0d781314f3d50897ef6cf377d70706d8a139639f7", "enabled": 1 } }, "livecd": { "100": { "checksum": "sha256:46a683c9c0693f57e0a687926c0ea32a8a176eaa168d1bf86e66463395b314cf", "enabled": 1 } }, "lldpad": { "100": { "checksum": "sha256:9fa26b1dd3ca8e52c93a1320fb66fceb0c8e83730227023d221a6a426b9dddf0", "enabled": 1 } }, "loadkeys": { "100": { "checksum": "sha256:8e876286cd84dd189d40234f691174bc5e0c214056ff0a9f454400cdafdcd6cd", "enabled": 1 } }, "locallogin": { "100": { "checksum": "sha256:f90a7ca564fcff2da3a9effa5c50e8c2a92de3e4c3831f741bc3f8cfc5be5808", "enabled": 1 } }, "lockdev": { "100": { "checksum": "sha256:d181bc52a0d2af9fe776462be417f9f28920a71f01f9894557e9bf7e128ddb5c", "enabled": 1 } }, "logadm": { "100": { "checksum": "sha256:c573637379d69beb37674b0ce9da5f57e60238a88ddad2faa1c3d456cd7df71a", "enabled": 1 } }, "logging": { "100": { "checksum": "sha256:d3494d1da872d6c128a1ab751a7ec84abb1ec73dccbfae200707935617fb8888", "enabled": 1 } }, "logrotate": { "100": { "checksum": "sha256:97dfef99f123dd0474194e78c4e8760df60d46db61f8460161bbcf4b78f7c530", "enabled": 1 } }, "logwatch": { "100": { "checksum": "sha256:c68409b41b76b32647af8f19d1f8f6ebd43430dc820e55d2b094730950a96419", "enabled": 1 } }, "lpd": { "100": { "checksum": "sha256:24594a8f8d0e864712707e8c4114f6d68dd19824045a9f3d10b8f590fc506858", "enabled": 1 } }, "lsm": { "100": { "checksum": "sha256:7df22b87a981a09999b90bb4ecea9713b595814a385fce77eb67e3c20042aa09", "enabled": 1 } }, "lttng-tools": { "100": { "checksum": "sha256:9f30b434e6c08d9a155bbc24ef6d71745d8d252c96d0ded0ccb6d158354c6546", "enabled": 1 } }, "lvm": { "100": { "checksum": "sha256:f0b2c006055d609858e59815fcd002fab6ce4a77761dfd2cc18b56219bf56fc1", "enabled": 1 } }, "mailman": { "100": { "checksum": "sha256:50901ef3e7ccc1828edc3fde37485af69afe28048ba25b7d9e3578b063d559d7", "enabled": 1 } }, "mailscanner": { "100": { "checksum": "sha256:914891424131853f76600695c57ec6283f9cc3c501c06e1427ef99f83431d1cd", "enabled": 1 } }, "man2html": { "100": { "checksum": "sha256:a7af8f4be03248a6d91c5df6c2becd22e3f45a71cecd671d44d2830fcd9f4a3b", "enabled": 1 } }, "mandb": { "100": { "checksum": "sha256:4afad481406029752a7960f97c253b41b8251018b219e16a858ccd029f9bb5f2", "enabled": 1 } }, "mcelog": { "100": { "checksum": "sha256:047b1501df33727c4257a655f1ad55aaed7da1d763782f7da45dfb1b19fa6a02", "enabled": 1 } }, "mediawiki": { "100": { "checksum": "sha256:7ce2990caacbf36b3ea6481def0630fa02e220ea9a4e2dc107251bce1368e569", "enabled": 1 } }, "memcached": { "100": { "checksum": "sha256:577518e3f4a348aaac570906781e3db46d9d53bf52d1a6f83c9f520dab8270e7", "enabled": 1 } }, "milter": { "100": { "checksum": "sha256:aa6812766cbdfad665abf5ad9633ffb945822d31714ffaa2bf4cb061f91e6b6c", "enabled": 1 } }, "minidlna": { "100": { "checksum": "sha256:f13c8b27b597e55d9e6342c756bc4d21de401fcbc4cb01e16c71c01a816c2e2b", "enabled": 1 } }, "minissdpd": { "100": { "checksum": "sha256:18649df6879516ea838eabbd19d6f7aefe2bb7da501c0b471c453d80ea4e02cb", "enabled": 1 } }, "miscfiles": { "100": { "checksum": "sha256:6f5792cde030e56b55f4b00cb53bd6d3724181c80aaaee5806e88c6a5298bb6c", "enabled": 1 } }, "mock": { "100": { "checksum": "sha256:3fdad88af11ebff86688d4db404640a1371af801e4a9b2f14a1b2abc1c76a503", "enabled": 1 } }, "modemmanager": { "100": { "checksum": "sha256:55370db1ae9a5e7ec0b4ec118f45ed61578656c6df3deb7b9e618362ddce4e29", "enabled": 1 } }, "modutils": { "100": { "checksum": "sha256:37b0e80e0d823ecd2606ece25c0b93eaea9f16da35261d79630b3e95c33a5fb4", "enabled": 1 } }, "mojomojo": { "100": { "checksum": "sha256:8ddaca2a54241adf5d42f9205cd073c7062e7692f31ebe15eb9e988e12efc9d4", "enabled": 1 } }, "mon_statd": { "100": { "checksum": "sha256:7e1bf20f24f949d86cf2cf1ecd59aef5aee94c9258d32a8fb4ee7646e4dc67e7", "enabled": 1 } }, "mongodb": { "100": { "checksum": "sha256:051c9eb43df88894edffecf5083ac5065a7051b743068b6f1866d923bca5ab95", "enabled": 1 } }, "motion": { "100": { "checksum": "sha256:154f6949eddb0f008a3b9166f736dde00bbc37df415f07c1263cd63ee609a5fd", "enabled": 1 } }, "mount": { "100": { "checksum": "sha256:acb9c37d19d0aec956cffe6f83d9593cb70bf8ad71ecf956e074b590286f489b", "enabled": 1 } }, "mozilla": { "100": { "checksum": "sha256:0a4982ad6ec48750689dc3c5d7ee946fca86c0c8b3763ee322aae33546109bed", "enabled": 1 } }, "mpd": { "100": { "checksum": "sha256:8807a444e6651ef4e67d0fed3bbc2558dc34b15e7293bfa03487d0a2c54ba905", "enabled": 1 } }, "mplayer": { "100": { "checksum": "sha256:3f999e549842f5a13f5f0d5aa4a752e74216c123f34f1797588b73d504afdc81", "enabled": 1 } }, "mptcpd": { "100": { "checksum": "sha256:f715709d773d8da3d99b4a7de877c7d7a8964d7a1342a61779edcdde2c75702d", "enabled": 1 } }, "mrtg": { "100": { "checksum": "sha256:740151127f1bcdb7699a6317eca571399e7cf58a755afd20dd141aec3035ab53", "enabled": 1 } }, "mta": { "100": { "checksum": "sha256:e04f6d7a528f9eeb7835428fe62d15043b9f4642cded56d682551e18920d04b3", "enabled": 1 } }, "munin": { "100": { "checksum": "sha256:1d2ea7b3403a2c521005ef980913298f01b2c00338366997273344590d0b218e", "enabled": 1 } }, "mysql": { "100": { "checksum": "sha256:07bc412265c70dcb7a3f85ba5265573ef76373364bbbd165ad3abb662e4de705", "enabled": 1 } }, "mythtv": { "100": { "checksum": "sha256:4cab8df79dc5a8e73ece06c418cb3ca58c74f669516b810de46b9f268d78ee8c", "enabled": 1 } }, "nagios": { "100": { "checksum": "sha256:712912fdf95fbf2adc1cb7427524d5536f0cd591bf3e77b63ef069213ae86704", "enabled": 1 } }, "namespace": { "100": { "checksum": "sha256:c032c4722dc3ec48bc9f196ae3de9ef9fd1c04554aa0913d99ade36ac0cf85af", "enabled": 1 } }, "ncftool": { "100": { "checksum": "sha256:e23fc9d15551fc1d7ba386350226d425282fcc090b9dfdcdb7065ebf3385666b", "enabled": 1 } }, "netlabel": { "100": { "checksum": "sha256:d9b7f3b8cb9b0ca564b43893085106dab4a6f6e7bee3143d778fc28218910435", "enabled": 1 } }, "netutils": { "100": { "checksum": "sha256:e828ed9b31086b0b53a5cf66cb2a580aae3e3290156b0d32c07eec5b1bd9cfe5", "enabled": 1 } }, "networkmanager": { "100": { "checksum": "sha256:bc61c1cee0f527dce7154fefb7906e6262e81d84108c7999f8db9cba7e4eb44c", "enabled": 1 } }, "ninfod": { "100": { "checksum": "sha256:4ed636e37871cc6a707bd29f6d990f4ff13d98bac2d8bf37922983d6289b3411", "enabled": 1 } }, "nis": { "100": { "checksum": "sha256:bce01b1750e64ca5f88e7d00d7d4dc88d5ebd430561f9b3cc680484b631c6c8a", "enabled": 1 } }, "nova": { "100": { "checksum": "sha256:16ebf87ef7e152147a0aa89e34161622d387348d4376f2eb91d971ef8ab476a0", "enabled": 1 } }, "nscd": { "100": { "checksum": "sha256:c7e5598679b39499dd00ac032b9a85b6e7e367bc64a0e2827cd799bb7e920a72", "enabled": 1 } }, "nsd": { "100": { "checksum": "sha256:75e019e0b2eda6179a3b39defbd43329dff4627e2c6a06c868ebee8b584332ee", "enabled": 1 } }, "nslcd": { "100": { "checksum": "sha256:69c6eaa2ef51784392d46e962c0f9b9439616dc97ba3105ff54e275a7b1e9db4", "enabled": 1 } }, "ntop": { "100": { "checksum": "sha256:cd4113e904b08b55f2a611ab02edbba284a91097a468b0d6b0d98a5dff397220", "enabled": 1 } }, "ntp": { "100": { "checksum": "sha256:5f188e0c5dabce04e39700cab0a3052cde82676b66e13b054d55f6f78f8d4b30", "enabled": 1 } }, "numad": { "100": { "checksum": "sha256:ed1193b412ae24e79c9a91531c99520b68ded7151bbf0f8cf0353d0dab713dec", "enabled": 1 } }, "nut": { "100": { "checksum": "sha256:54c04020d40b310f4b85f747395b3183859d816a6f181e4a54f5bd6f23b660df", "enabled": 1 } }, "nvme_stas": { "100": { "checksum": "sha256:1098f37f7b13bb6fa26d45bd4c54fd8cb38eddd1b4d02efaf88241f7338022c2", "enabled": 1 } }, "nx": { "100": { "checksum": "sha256:a8b41cc8424e4d9d67310c486221a9fc2701a9508104bdeb076f2ece6c94957c", "enabled": 1 } }, "obex": { "100": { "checksum": "sha256:4810ca21adc14e1f8b8f2a5b91e628038786e71ff84a2b3d701cdcafc6baf16c", "enabled": 1 } }, "oddjob": { "100": { "checksum": "sha256:d09199963de83937d060d3199ad48bf9f6753507c2748ad11d86628469f1e5b5", "enabled": 1 } }, "opafm": { "100": { "checksum": "sha256:020fbfc8bc4b762f26e66c399e2ae1f202c409d7ae1f10ef84f94c9672a1e55f", "enabled": 1 } }, "openct": { "100": { "checksum": "sha256:fd4ba492c3bc1b35ee1d9e4cf2b1fa21ac80ed31c9c76cb4503996f7c6367898", "enabled": 1 } }, "opendnssec": { "100": { "checksum": "sha256:cd5781c1d95be89112ad12c0d1341cc43b60259d4eea9b11442bc19da8ee462c", "enabled": 1 } }, "openfortivpn": { "100": { "checksum": "sha256:9a125b8b0328b3f01aea6aa395ce5117bc3a683e2b45b12e69dc6a95116ce203", "enabled": 1 } }, "openhpid": { "100": { "checksum": "sha256:65ec5e445fa5dd1471b83bccc319e9403a1c5888423442382751b1bf91c5188d", "enabled": 1 } }, "openshift": { "100": { "checksum": "sha256:07bea1192b1c51dd443db5a17d844ba3d7383abc3a3e7e979dcf0143753af074", "enabled": 1 } }, "openshift-origin": { "100": { "checksum": "sha256:88873289a3c1bd0c0c5d636b689bd8e1fe194ca528720c3191fb61b2ca024b50", "enabled": 1 } }, "opensm": { "100": { "checksum": "sha256:76a9f4d16f6b8fe2cb94b30b0c557dd4b067a4cb49502ff812b95290539062f2", "enabled": 1 } }, "openvpn": { "100": { "checksum": "sha256:ec0e41fdd7c04028cedf656a3fd60533adc778663a70734f5f2d569f5e79f53b", "enabled": 1 } }, "openvswitch": { "100": { "checksum": "sha256:1ab8ca3123bd0bbfbe2baac2e44124259401557c590cb8e906b06f28a4b9e32d", "enabled": 1 } }, "openwsman": { "100": { "checksum": "sha256:417b01a65501649535510d72e5fca7174183b5c873c1ddcb6921c052c8be3fc3", "enabled": 1 } }, "oracleasm": { "100": { "checksum": "sha256:6070bb82ad7b693e1ce70bcec6b975c4f5baec8a51e879f27e2c80ff2e076755", "enabled": 1 } }, "osad": { "100": { "checksum": "sha256:7bd10b072510204cc5f32e694b6eb1389fc20c40186fad86b76079e774023f4b", "enabled": 1 } }, "pads": { "100": { "checksum": "sha256:aeae1d20fad2ecc5e2f3ac84829becd50d70a7c41e8183c5b817e0d3f97f9484", "enabled": 1 } }, "passenger": { "100": { "checksum": "sha256:93cb4d1c0096f51325fa10c42ab72db38b5974c42b7d9fa6789482f56ce688c6", "enabled": 1 } }, "pcm": { "100": { "checksum": "sha256:4bb8d2c73a705f3051ce26155740950675bb556b11e2a7f59bb390ebbf7f9a87", "enabled": 1 } }, "pcmcia": { "100": { "checksum": "sha256:722b33d72298e2c8fa2900f1592c92673d098fcb6cd5802449086c918ca104dd", "enabled": 1 } }, "pcp": { "100": { "checksum": "sha256:a34e540529562d554136e3710fe67e8ca6e9ccee02b76dab369a448b58236f34", "enabled": 1 } }, "pcscd": { "100": { "checksum": "sha256:f50088e07d5fd52c3ad752e30852f0d3085cae9f13d2da6fd8c8141f1568ca28", "enabled": 1 } }, "pdns": { "100": { "checksum": "sha256:2d20aff28dd9cac707097c581edf455d754629927a32cc07854eb9baedb23b0a", "enabled": 1 } }, "pegasus": { "100": { "checksum": "sha256:81bedf6564a2a1e2b7ef03a9e5f646184f770202ad403e031f430e494abf8e88", "enabled": 1 } }, "permissivedomains": { "100": { "checksum": "sha256:2453bad4ace526f3cf2c60b358e95a5476692ef25da107b10f52f3af27c056d2", "enabled": 1 } }, "pesign": { "100": { "checksum": "sha256:544140731f1c09b504a6b85ec6ffbdda47c549c9858d6c2264498bfc56fd6127", "enabled": 1 } }, "pingd": { "100": { "checksum": "sha256:a4f1dd19ad8877fb1b48c431790bc350af44d3e2d0d08d59497d2a4425ce2d80", "enabled": 1 } }, "pkcs": { "100": { "checksum": "sha256:c0fe90911ac32abd43e43d9e25dd5a15c85301a1ad0ee579fd3997b9c25ad201", "enabled": 1 } }, "pki": { "100": { "checksum": "sha256:16249004873cf9214f949f73481b5476c0ab03e349542512af9b57aeb637c047", "enabled": 1 } }, "plymouthd": { "100": { "checksum": "sha256:4e33d3ebd3e3cd1fc439ca7bcf3c11086799c7af6a1e12460eb4fc5bab23b0be", "enabled": 1 } }, "podsleuth": { "100": { "checksum": "sha256:7f525e710a117aeaec78a9ab1cbbdd54a150ffe26ed5a1bc29595824fd9e1931", "enabled": 1 } }, "policykit": { "100": { "checksum": "sha256:fe85c9803d13edb28fd5a47803918af2a1e70d66b57e920fefe8e6cfdf778171", "enabled": 1 } }, "polipo": { "100": { "checksum": "sha256:8ab8b8f9438c6ae6c96323b6cf5937528d8223c38629f9bed4a2df00f4b9b686", "enabled": 1 } }, "portmap": { "100": { "checksum": "sha256:59c39477c07ba01841c150a2232e21074575ad3f93782e73b35da8f7acb121af", "enabled": 1 } }, "portreserve": { "100": { "checksum": "sha256:00aa972aa3ea3cf49a0134ebc75df0d6b7eb87ca7b31c61b3107001111abda8a", "enabled": 1 } }, "postfix": { "100": { "checksum": "sha256:68f178fd86c386a5d4df236f983239aa04eb6482304df3e9eb56340b744f5d8c", "enabled": 1 } }, "postgresql": { "100": { "checksum": "sha256:bab7ecb38d7aa40fc1de9d3f1a06e0f83c473e22e2f2391cbdc50e27ef92f7dd", "enabled": 1 } }, "postgrey": { "100": { "checksum": "sha256:8e1915326802644782f3041e65e3b1dfb6f534f2a66f4cb00ca8391f35ef2048", "enabled": 1 } }, "ppp": { "100": { "checksum": "sha256:912c8757c9d3cc46657aa135743c58836db26f4a6abf17807017c86a8512d90f", "enabled": 1 } }, "prelink": { "100": { "checksum": "sha256:0b0ad935c694badc8da9368226c53640cfade72c6828d094924029340ad5521e", "enabled": 1 } }, "prelude": { "100": { "checksum": "sha256:92c5efa0190469791dd3a105bebd81b2891596683001c6cb0d7446fc0d5125a2", "enabled": 1 } }, "privoxy": { "100": { "checksum": "sha256:0c0a1216baa5adcd33f85d92f248893c11cb55b5b068db990b2f94d9d7914932", "enabled": 1 } }, "procmail": { "100": { "checksum": "sha256:ab3962f174f9a9a125e2aad039584a83294b11d3ee56c8b477217a8193afc87d", "enabled": 1 } }, "prosody": { "100": { "checksum": "sha256:3573621c7a830f0d511041287fe8bb073eaa5e200c5b8b2e72096bf469c35f1d", "enabled": 1 } }, "psad": { "100": { "checksum": "sha256:72c77f99e96935aa093669bb095f914484b67e54ae66b4ef8ca780d2758cc6c8", "enabled": 1 } }, "ptchown": { "100": { "checksum": "sha256:318c62c2f63baefa43a61f8ef58bc168596087cd892a63afee72b1cb29cac0c7", "enabled": 1 } }, "pulseaudio": { "100": { "checksum": "sha256:3a7087d4680d1a35893caebf703937f36c1b1799dabc446b02428afc6c163c6f", "enabled": 1 } }, "puppet": { "100": { "checksum": "sha256:366806c68d03db1138cc9a3a08e0fdc4883825c20d1bed9a8e64ce7662bfb032", "enabled": 1 } }, "pwauth": { "100": { "checksum": "sha256:d23c293e39c1a025d375b714dd9319d27545c6c62b58342ef530763b84a24b16", "enabled": 1 } }, "qatlib": { "100": { "checksum": "sha256:cef00aa0bebb89745286c7db05374eb06fe9637c480d295f80f31d21be53672f", "enabled": 1 } }, "qmail": { "100": { "checksum": "sha256:5617289b46ae349fafd315cf92082d64119c6a0b5606bf79aa17fc99e34985b6", "enabled": 1 } }, "qpid": { "100": { "checksum": "sha256:1ee83ce320b057b0569c4bc9f8abdab6233eaf6fcd9c1a78975040bb2d9ddc56", "enabled": 1 } }, "quantum": { "100": { "checksum": "sha256:704e2c7d5fca4cb84e320b9359c6168037ebc87908330c0053c143c8821d3863", "enabled": 1 } }, "quota": { "100": { "checksum": "sha256:58dca8ea36c665ab74eb79cd51d54a7224514660bdfc4b666c2b8031e2877da3", "enabled": 1 } }, "rabbitmq": { "100": { "checksum": "sha256:acc51303259d5e6049265839c5ed13703baa80ae3477ae292103b4665c83dfbb", "enabled": 1 } }, "radius": { "100": { "checksum": "sha256:b53cd6e9f7493004e2b5bdf30e368fbc90b4f971db3fe56e55f31cc839969237", "enabled": 1 } }, "radvd": { "100": { "checksum": "sha256:c909b5d3b8ba285d71358dc7c9fca270e7b1586266c597a60b167ff600d10f7b", "enabled": 1 } }, "raid": { "100": { "checksum": "sha256:071d5ed18ad1c7d334bfba17d392c80a588ad00b1052ab0e21f9c53b0db95e4f", "enabled": 1 } }, "rasdaemon": { "100": { "checksum": "sha256:6d5c3ce954d1cef655ba323f1fef77bda8813815c0d991d020e0c760a4089aa5", "enabled": 1 } }, "rdisc": { "100": { "checksum": "sha256:3b10e49a0fec5da0e8998d77835464b25ced6ee9bd04a24dd861601a9aa02d2d", "enabled": 1 } }, "readahead": { "100": { "checksum": "sha256:e5b3d0b401606c60c1acdfce14c7b6b3f3363a1f58d8d91dd3be36e3df38005f", "enabled": 1 } }, "realmd": { "100": { "checksum": "sha256:b37f4de6bcd25f24eee675c18e2f1842023f0cf9a9a9a7923b64f7eaa875494f", "enabled": 1 } }, "redis": { "100": { "checksum": "sha256:b862f723b5285ad797d5a7546a17b1080690d9218a23b84bb0f7fe618169610f", "enabled": 1 } }, "remotelogin": { "100": { "checksum": "sha256:622d5ae5349b504c7ec9e472a86c5a1e1bc27606868d53d154112a63299c8dc9", "enabled": 1 } }, "restraint": { "400": { "checksum": "sha256:95d0d03fbc1d4147f02a0b3da7cd76efbdd75d1f5812cf6160e922336abbf270", "enabled": 1 } }, "rhcd": { "100": { "checksum": "sha256:fcd8c8415a58d937e726318252127c28e69af2819be01b977b4dda03994834a0", "enabled": 1 } }, "rhcs": { "100": { "checksum": "sha256:1e8298e1e453be52e1a86e391b827f2c71729c066f7d6c63abe7843e28647207", "enabled": 1 } }, "rhev": { "100": { "checksum": "sha256:eaa7e5657902e356ab7700656ff4c185eaf8df675d0a9a3f9ea1a4d8d8fb577d", "enabled": 1 } }, "rhgb": { "100": { "checksum": "sha256:5271469985f35fb7911a72e9e705bc35971294386838f1f11033657874821cea", "enabled": 1 } }, "rhnsd": { "100": { "checksum": "sha256:2a502ce16b7c56df48d0798ed7ca60370d0993ee425a542bf0621b51d4caf2d4", "enabled": 1 } }, "rhsmcertd": { "100": { "checksum": "sha256:80b01c4c3ddf665dd2fd5391b6069e9b868573f7c7a9d4421bc5bfeaadc71a22", "enabled": 1 } }, "rhts": { "400": { "checksum": "sha256:39f15722a115b11064c20b8bc2758e8fe06531a8f923ea00a3e7079a0554e3d6", "enabled": 1 } }, "ricci": { "100": { "checksum": "sha256:dd355c8a0a048baee894e0951365b525bbf5d7cce5907c9919bb0fffb1876661", "enabled": 1 } }, "rkhunter": { "100": { "checksum": "sha256:95ecb10dfe02e0384b0f220c46a7d0c85166a4e2cca0bb55654420b73611a646", "enabled": 1 } }, "rlogin": { "100": { "checksum": "sha256:d8594a73f246468d42569485bb148cbf1b582779074735f6da46e92f763984e1", "enabled": 1 } }, "rngd": { "100": { "checksum": "sha256:50b57a4776d2c6de8b1698bd31ecbfcc98d90d4a9bd2c95672b114c9637ce940", "enabled": 1 } }, "roundup": { "100": { "checksum": "sha256:54c5e85363d326f78c5896bf87c36fca9bc28acee3bd9da4d6d22f090dc9f7c6", "enabled": 1 } }, "rpc": { "100": { "checksum": "sha256:404f69ae03180ddce44c15efc636e5d7ff34b2108cb2c09624075be42a0545b1", "enabled": 1 } }, "rpcbind": { "100": { "checksum": "sha256:f5c703b4107b2616b1adf105b5c592c79f03df1f3a24f31453a242cdec50ea97", "enabled": 1 } }, "rpm": { "100": { "checksum": "sha256:bb52185531861597a0b44bca75497c08ac4c07e10ec63505c333bde800176fb7", "enabled": 1 } }, "rrdcached": { "100": { "checksum": "sha256:2cb11286a0afa94eccbd33de5a32b16d9e178e41ac1fec9dcd801b77dd257051", "enabled": 1 } }, "rshd": { "100": { "checksum": "sha256:0052600f999e54f224d1210c5920b42227cd98ce8db24155b0cfecdf4f6c6474", "enabled": 1 } }, "rshim": { "100": { "checksum": "sha256:7825528bbcdb081ead1d0d2da0d5156e430f93d003f58dcebbb8997c906f0dec", "enabled": 1 } }, "rssh": { "100": { "checksum": "sha256:cde9bfc64b345835e9f8012426cf218bad9972c70e22e0ef6d6a90cc5fa7163c", "enabled": 1 } }, "rsync": { "100": { "checksum": "sha256:1295d1a5434aa1df7fd0f0e2a9b2b045b0ec51bd8619d528349530b322eddf24", "enabled": 1 } }, "rtas": { "100": { "checksum": "sha256:8c5c3a495b42ed14f56e360dd9eaaa0df2426f2c93676880f129bab95e247338", "enabled": 1 } }, "rtkit": { "100": { "checksum": "sha256:7b0c8b0fd23723aed98bc581f2ead99da87893c6ba042ec19ca731e1d802eab6", "enabled": 1 } }, "rwho": { "100": { "checksum": "sha256:fe02b4fb75b8a75ac77b53ae13939199422543f68e71829fd1530aa34c8715cc", "enabled": 1 } }, "samba": { "100": { "checksum": "sha256:b0af5e99697f8e864538813653ea879ea50b6a9269424ca1deaa95ec1a4a7778", "enabled": 1 } }, "sambagui": { "100": { "checksum": "sha256:cde09885f6d657107c517f09f01933816199f0694799fe3df4ac281b411edbb9", "enabled": 1 } }, "sandboxX": { "100": { "checksum": "sha256:ffd79700483dc635b0396df2f9ca75062aff985e7874c2014bca95e80b08bd5d", "enabled": 1 } }, "sanlock": { "100": { "checksum": "sha256:fcd9cea4d61e3597a5e5cb1bd36efc228a21e5b8ee79c68b44e8a93aae834fb6", "enabled": 1 } }, "sap": { "100": { "checksum": "sha256:4cffd87a27c2b48cc59f596a42e860ea52acdfcbff0c2a1e4aeafa3338b29e6e", "enabled": 1 } }, "sasl": { "100": { "checksum": "sha256:ed40cd297b033fd86770692e7d3ab76598684f521f079475d8b44953116291cf", "enabled": 1 } }, "sbd": { "100": { "checksum": "sha256:8c40daa2d8288e7535e9db7724d16a34add080c65fab32bdf5e10cdace95a063", "enabled": 1 } }, "sblim": { "100": { "checksum": "sha256:73b934662023f728bfeb5cac500547e15e389bad3575a3750bf05901d4396531", "enabled": 1 } }, "screen": { "100": { "checksum": "sha256:8dcba8c325fdb2c56f7c7c77362e287469484b19f0b491dd7bc0abc6e033378b", "enabled": 1 } }, "secadm": { "100": { "checksum": "sha256:2d9e07e5ee506dbf6cf976e257210fdb14222c1e437b1f7dceb1548fa399beba", "enabled": 1 } }, "sectoolm": { "100": { "checksum": "sha256:e6c4cabc853877f459381f06d4d47660ce9360a188ab87bb825f1f1eaed39d7d", "enabled": 1 } }, "selinuxutil": { "100": { "checksum": "sha256:38da361f759758e151b7f8612954c91c0fde09615e1c82242c80706231e16355", "enabled": 1 } }, "sendmail": { "100": { "checksum": "sha256:286d18507d80fcfa7cb83ed21321df23c3a29001777890137c1b52c113cd8489", "enabled": 1 } }, "sensord": { "100": { "checksum": "sha256:6681fdfffe9ba81fa3712f665b98feb68f9331f8f282faf19a4ae5e70b02bd63", "enabled": 1 } }, "setrans": { "100": { "checksum": "sha256:69a3bce8a9e35afa4c3b4c7e2f4ba604491313a688d36e3158f96e90343ea370", "enabled": 1 } }, "setroubleshoot": { "100": { "checksum": "sha256:c0d9c0c6cea100697cbc684f2ab74ec66d7c121ca88200752585aeaf4c2b2596", "enabled": 1 } }, "seunshare": { "100": { "checksum": "sha256:219912d16220957b98bbe694f06c3509d2f884ff8418da1eeeda43f4dce81dc9", "enabled": 1 } }, "shorewall": { "100": { "checksum": "sha256:d0dd25c1a5b77a28fcf61f4796c479a026a8e2ec944f5e4d1492cd552aa9ca02", "enabled": 1 } }, "slocate": { "100": { "checksum": "sha256:8d149bca3982ac8c0f7376571d1f0614f05eb2fe7465951eefb17153f642ee40", "enabled": 1 } }, "slpd": { "100": { "checksum": "sha256:6798797c2dda6b8b24aada5a4bbc4f8775faca22435ed1de39193457043f8d8b", "enabled": 1 } }, "smartmon": { "100": { "checksum": "sha256:9de88cb8f60049e56dc5ead706a324189ff276929ccc3bf4ba7d06e96c0d8414", "enabled": 1 } }, "smokeping": { "100": { "checksum": "sha256:29560964ba6e3cb5fc392417ff4aa3e50a5f15c2bc21bdf37088c87868ffeec7", "enabled": 1 } }, "smoltclient": { "100": { "checksum": "sha256:0e47cf6e775a1e2ecb506e444731e086d057b7f2c2aeb6bce7b636372de88d95", "enabled": 1 } }, "snapper": { "100": { "checksum": "sha256:e264d81b3e1173c98c1efc90529d7f9c2c9047b843e29a4fb807d0b1b96b4685", "enabled": 1 } }, "snmp": { "100": { "checksum": "sha256:ec6638b69ec04510a807a5d2c97568a670f80a7999ee711e5e1874136923718e", "enabled": 1 } }, "snort": { "100": { "checksum": "sha256:67c1751c124c249fa15c36f948b60a76db3729f819678cecba569c79abec3af2", "enabled": 1 } }, "sosreport": { "100": { "checksum": "sha256:345531e273b1a0c0d6e02a1898d60759007d5126ec2016eea320832c33ac8211", "enabled": 1 } }, "soundserver": { "100": { "checksum": "sha256:932d162b8365f213b97eb6a7e62d20b8c30bfe1863b6534462da96077d6347fc", "enabled": 1 } }, "spamassassin": { "100": { "checksum": "sha256:fbceb12b459776c22867efa6bf2d07657e058c4a80d6f20e724086ec58e600b0", "enabled": 1 } }, "speech-dispatcher": { "100": { "checksum": "sha256:092593b379ff3e7b3d15b0dfdd2adfde73fb76401d1b6b6e79a4efccbf03f0c9", "enabled": 1 } }, "squid": { "100": { "checksum": "sha256:4374a853cb79bd24d05f65585ad1a74a7b9da4d3454ca233da351e009e75ed79", "enabled": 1 } }, "ssh": { "100": { "checksum": "sha256:493ee98f9cda10ca995eb5600d2940556d70cb3f544501c6bd1968745cc234fa", "enabled": 1 } }, "sslh": { "100": { "checksum": "sha256:5dc03f6d33a6caa7c2fb5ede0f8ded4ecb8eb244443614189cb5909521cab539", "enabled": 1 } }, "sssd": { "100": { "checksum": "sha256:4ee157395107db3b146239f14c9868f6e9d71199520e5d2e473e5e836d0635d1", "enabled": 1 } }, "staff": { "100": { "checksum": "sha256:6a89d52653ce210b1dfb2d2b10e184cacb18976d238fc92b349c40661fdcbf47", "enabled": 1 } }, "stalld": { "100": { "checksum": "sha256:14aa28921885421f92fb51920ca0d8fde122881ac9e28d024c80a7b234b82dd3", "enabled": 1 } }, "stapserver": { "100": { "checksum": "sha256:1610ffc61426c0e791bd7a6a2628935edf3d66fc7818fb42c314044a26ac5d2d", "enabled": 1 } }, "stratisd": { "100": { "checksum": "sha256:5025b28ad7090d4802078c92ba257b46909bbb74fcd6cae0f2a5d65978a5ff8e", "enabled": 1 } }, "stunnel": { "100": { "checksum": "sha256:67b1c960e0cd11cdf1446615cf633700585e69f41c4039cfb183d69527648f1e", "enabled": 1 } }, "su": { "100": { "checksum": "sha256:60850779bd04d036ea792653d4e794f0e865936648c03030515d8125cf412b4b", "enabled": 1 } }, "sudo": { "100": { "checksum": "sha256:0965d3f30b2813a4364e9ce8da0f8916579319ed9507640383c3bf195ae1c00e", "enabled": 1 } }, "svnserve": { "100": { "checksum": "sha256:d589d2e8bddbcea03272bb589cffde776ccb410ac685e5071896fa8e59660c7f", "enabled": 1 } }, "swift": { "100": { "checksum": "sha256:407ddd533381c1f822ec8471e639404c4b926e7e635b3d25d4ba1403a5c517b3", "enabled": 1 } }, "sysadm": { "100": { "checksum": "sha256:1d310d8df59d431579fa3568b7ac14f853223aadd4087fe4c57a1add6606987d", "enabled": 1 } }, "sysadm_secadm": { "100": { "checksum": "sha256:6a0a3b3c4028f304beeacb7d06c62975a2bfb13d67a6a4fe6b95f6d6eeaa74b7", "enabled": 1 } }, "sysnetwork": { "100": { "checksum": "sha256:56b999cfaf7defa2c5cdd5c0c16e7c9c1de940dcc528778ff752fca51f95b38a", "enabled": 1 } }, "sysstat": { "100": { "checksum": "sha256:358b9c77d350c4bc17cf773e99f81d5267f8c41d47402234be37dd9f31655ad0", "enabled": 1 } }, "systemd": { "100": { "checksum": "sha256:124460467ec0f3f7443a67583ef4613154b2b92615e0d779a3ec8060c5a8ce4a", "enabled": 1 } }, "tangd": { "100": { "checksum": "sha256:a2e2a5d807eabe30a09085dcec38d530781c1f02e840a6ed02ccc32bd2738e60", "enabled": 1 } }, "targetd": { "100": { "checksum": "sha256:4a9e67698b4759a406188562c375ecd847f205c38873fb895c3c20abc8bac869", "enabled": 1 } }, "tcpd": { "100": { "checksum": "sha256:b69263d84bf511825be889b667401c30964c7d30d1147d577ae294cd7f05aa01", "enabled": 1 } }, "tcsd": { "100": { "checksum": "sha256:fa40e623b09fb5f61c2c2113a8ed0d17a4a5d105b9f22389ce088a766c0042c0", "enabled": 1 } }, "telepathy": { "100": { "checksum": "sha256:d98021b3bbdd6255c2009e5adafbf41faf3370434396cf2e928e7df693dd50b6", "enabled": 1 } }, "telnet": { "100": { "checksum": "sha256:abf90c09c67553f2b04cab7461138bfd8689752a752a79e36e248f209eae73e6", "enabled": 1 } }, "tftp": { "100": { "checksum": "sha256:f5dcde99c32caa66ea23a9be3b2d6d95550eff7db62b11733f16f1f5eb01bce2", "enabled": 1 } }, "tgtd": { "100": { "checksum": "sha256:6a96f6d7abaa0e1951832347e924c5aff5c66194da3881ec4855e1401a5ce6d2", "enabled": 1 } }, "thin": { "100": { "checksum": "sha256:b0c1055c6a8419035ee9a9a45de2bcdd7183a906bb40d98bb2be80b882a60c51", "enabled": 1 } }, "thumb": { "100": { "checksum": "sha256:928596446716a5d1ef1f8958d150959e59a6258975ebf453c2d5a2747e6d59fd", "enabled": 1 } }, "tlp": { "100": { "checksum": "sha256:540f0a0339feec1e11d145c6ba99f700750451a2fad43190e256e590c3835950", "enabled": 1 } }, "tmpreaper": { "100": { "checksum": "sha256:4370bc58da3c779088151878e7291e03fcd2ed706425b4f2eeeed7097f8adcad", "enabled": 1 } }, "tomcat": { "100": { "checksum": "sha256:98e40c398d94d82eea102055d52f59f77581955b0e463dd95386b4559a4cc852", "enabled": 1 } }, "tor": { "100": { "checksum": "sha256:b309ee2c391d91ace45662d2152870961635ca8968487d0f2570508e160dbd9b", "enabled": 1 } }, "tuned": { "100": { "checksum": "sha256:90bcbb5cad47c032a02390e0c4b22ff79f93fbfd2eb4a385441a87532d6aa632", "enabled": 1 } }, "tvtime": { "100": { "checksum": "sha256:a4438b3063cc1760c663e35a7be70a7dcd099de505ce1562e02090ac6b635787", "enabled": 1 } }, "udev": { "100": { "checksum": "sha256:f308ba2a718171dc871e2586f35a90cb1bbc37decf13e714af72b38cc44505a2", "enabled": 1 } }, "ulogd": { "100": { "checksum": "sha256:05aa3f423e0505dfef9b5e311b0f0a71d93159c7f6440979d3c00a5d9d22ef4c", "enabled": 1 } }, "uml": { "100": { "checksum": "sha256:e0257dd4e250c5aa0baa8151d48942735949bd3018318c11c28a4a5df2aacb90", "enabled": 1 } }, "unconfined": { "100": { "checksum": "sha256:26c4304a314b04f6269b45b4285b449caad98d22de7b3017ef648212c6674da1", "enabled": 1 } }, "unconfineduser": { "100": { "checksum": "sha256:af2778eed6f3cc25c5390e4b1a892995c37a83f5952f51820007b8cd387833b4", "enabled": 1 } }, "unlabelednet": { "100": { "checksum": "sha256:c1de2adfab2cd577a0cc9f2c1a50cd6321324d31b9c3f81be9231f85d4acbbd3", "enabled": 1 } }, "unprivuser": { "100": { "checksum": "sha256:dc1ddbf10a94d55c476ff343c862c24dc649d82190a16a0fd924c527743a667d", "enabled": 1 } }, "updfstab": { "100": { "checksum": "sha256:11eb931e80afefafd0f1c41ea19add0c746170447c76dd629146c43bd9e98067", "enabled": 1 } }, "usbmodules": { "100": { "checksum": "sha256:3a8334855e0cfe67e2c29198baacf61ea0e397d5da30f60fb844044cd105557e", "enabled": 1 } }, "usbmuxd": { "100": { "checksum": "sha256:e883ad3abf2e2792abf4122f99db904139b2f1833b7d99c45b34a7c8244d7c66", "enabled": 1 } }, "userdomain": { "100": { "checksum": "sha256:b6cc74d18c0d17741092aaa8e830f160021ad38bdc38880e3b1488ef30f59a49", "enabled": 1 } }, "userhelper": { "100": { "checksum": "sha256:0a80be6b1635de7593f46421219be5a6ccb9c22973c60b74436ba16884d789fc", "enabled": 1 } }, "usermanage": { "100": { "checksum": "sha256:c56dfa32f163d476dd5761e762ea743189a97636acaf2b4eefcf88812e624877", "enabled": 1 } }, "usernetctl": { "100": { "checksum": "sha256:9c0430aca661392960008075f0a1a934f0c05239976e73c4f0408e9cb5e279ca", "enabled": 1 } }, "uucp": { "100": { "checksum": "sha256:527db920542a250b24e24cb8c6bd86502c3be7e270d11f43d4ecb1a6632f8421", "enabled": 1 } }, "uuidd": { "100": { "checksum": "sha256:b3df65ab2d82cbb05e872f00f46156cfbae937a3a6f850d05f43f4866b772af8", "enabled": 1 } }, "varnishd": { "100": { "checksum": "sha256:9174c3c68bf6b4c2b2b9be7e2a0d985ce6edfdec0dc4c225b427b986449d3e7a", "enabled": 1 } }, "vdagent": { "100": { "checksum": "sha256:52636260af5ae220698d505c6a90a33a71f5878a2cea3a303033c896f000b732", "enabled": 1 } }, "vhostmd": { "100": { "checksum": "sha256:df2d64343b871c840603e82f5fc1f7928c0345b314450e536f9ef66becf131ad", "enabled": 1 } }, "virt": { "100": { "checksum": "sha256:c209f246ed99c66defe424c0f1bd13e970df31d84d2f4e4a2a79ef1c00c3bbe2", "enabled": 1 } }, "virt_supplementary": { "100": { "checksum": "sha256:e8147b1556404b4a2ce521cc4c22e28d4ee42045ea2480043fc8bd2d5a593207", "enabled": 1 } }, "vlock": { "100": { "checksum": "sha256:54a1b28cebc1928622edc4f1f1762a5f2d57783be153052d592f2d10dd9b24a1", "enabled": 1 } }, "vmtools": { "100": { "checksum": "sha256:323b80c9ba1efb9852725d635a5aa442ae2bbd4ec74360cf75853616401dd4f9", "enabled": 1 } }, "vmware": { "100": { "checksum": "sha256:b68b94c1c4662f7005faac011c887619dbb76d69d3d2b63f354d48a144d0d63c", "enabled": 1 } }, "vnstatd": { "100": { "checksum": "sha256:c544da94ee903d411f7ab352b3fe286fa0a1b0fb8490d508565df66d5501b852", "enabled": 1 } }, "vpn": { "100": { "checksum": "sha256:035f214c319dcee2ce4e9751fa9cf1343d81d569409f3a56faf79e506e445504", "enabled": 1 } }, "w3c": { "100": { "checksum": "sha256:d6c6c226dcc92d9a668f42b86dde398d2d875b479915af656734357e20ed0e96", "enabled": 1 } }, "watchdog": { "100": { "checksum": "sha256:a237aeaf94bed19de41ef5289be96ce5bf85dce6f2fc4f9d36d6185b4e59520a", "enabled": 1 } }, "wdmd": { "100": { "checksum": "sha256:1012fdefa5dcf2e540f536de5da79df0ba715ccc2a86a414e2e846675d32495f", "enabled": 1 } }, "webadm": { "100": { "checksum": "sha256:f374fcd23ed5f30ce1159470af1e1254e0352a6ad7cf261287de8ec51f2b2626", "enabled": 1 } }, "webalizer": { "100": { "checksum": "sha256:b66dfdfd23ed304d5f416bb0d8de4565d28cd732253fc2f9e4b4bac59515790f", "enabled": 1 } }, "wine": { "100": { "checksum": "sha256:a4a3be2a0feea4de24ca26c0eed58c7025304d810ee9d4a3e1b7e964a8d28c6d", "enabled": 1 } }, "wireguard": { "100": { "checksum": "sha256:3084b8fd20df45b718cb6779240f4292a0437f536bde7db3616c40b1500ba0f6", "enabled": 1 } }, "wireshark": { "100": { "checksum": "sha256:3ec8e55883a8b73de513f533ed80b1b665e3ed650fdf2b471e5565e1cc3034d9", "enabled": 1 } }, "xen": { "100": { "checksum": "sha256:adabf99dccce9182074a23fb8f05ac537899170fbbfe022f39cb520275339186", "enabled": 1 } }, "xguest": { "100": { "checksum": "sha256:8a7f7d9514fa5257088e451dcf96c6c221a7b6499cae9ac15a688cbe2e84def9", "enabled": 1 } }, "xserver": { "100": { "checksum": "sha256:ca8b72b391320d267580cebf2ad9ef811a781b34647e1c63e525e982911cf956", "enabled": 1 } }, "zabbix": { "100": { "checksum": "sha256:ed9800f41440cf137baa7864acdd8a503b7e051cf023d252b87daeb5357be889", "enabled": 1 } }, "zarafa": { "100": { "checksum": "sha256:c3a9d1c95f377fafb9e94a84bec89645a1053d23b557b143367992eae9993596", "enabled": 1 } }, "zebra": { "100": { "checksum": "sha256:77d57a56a3196f09ea5eb808f8c3b4cef333b241e9007b91779ba78a857e5cfa", "enabled": 1 } }, "zoneminder": { "100": { "checksum": "sha256:675bdf525f863eaf17fe6c25507cf2dc3c4ab05d1e4b0d3477573d34980a5cd2", "enabled": 1 } }, "zosremote": { "100": { "checksum": "sha256:79bb5490f9f16e8caff67fc87a834bf6f11647cfed990c421616eb1adaac5c70", "enabled": 1 } } }, "selinux_priorities": true }, "changed": false } TASK [fedora.linux_system_roles.selinux : Load SELinux modules] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:115 Saturday 11 January 2025 11:30:09 -0500 (0:00:03.011) 0:01:34.577 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_modules is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Restore SELinux labels on filesystem tree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:128 Saturday 11 January 2025 11:30:09 -0500 (0:00:00.044) 0:01:34.621 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Restore SELinux labels on filesystem tree in check mode] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:136 Saturday 11 January 2025 11:30:09 -0500 (0:00:00.039) 0:01:34.660 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:153 Saturday 11 January 2025 11:30:09 -0500 (0:00:00.063) 0:01:34.724 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:157 Saturday 11 January 2025 11:30:09 -0500 (0:00:00.061) 0:01:34.785 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:166 Saturday 11 January 2025 11:30:09 -0500 (0:00:00.105) 0:01:34.891 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:175 Saturday 11 January 2025 11:30:09 -0500 (0:00:00.042) 0:01:34.933 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:182 Saturday 11 January 2025 11:30:09 -0500 (0:00:00.038) 0:01:34.972 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml for managed-node2 => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:14 Saturday 11 January 2025 11:30:10 -0500 (0:00:00.134) 0:01:35.107 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_spec": { "state": "absent" }, "__podman_kube_str": "apiVersion: v1\nkind: Pod\nmetadata:\n labels:\n app: test\n io.containers.autoupdate: registry\n name: nopull\nspec:\n containers:\n - name: nopull\n image: quay.io/libpod/testimage:20210610\n" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:21 Saturday 11 January 2025 11:30:10 -0500 (0:00:00.059) 0:01:35.166 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_kube": { "apiVersion": "v1", "kind": "Pod", "metadata": { "labels": { "app": "test", "io.containers.autoupdate": "registry" }, "name": "nopull" }, "spec": { "containers": [ { "image": "quay.io/libpod/testimage:20210610", "name": "nopull" } ] } }, "__podman_kube_file": "", "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:33 Saturday 11 January 2025 11:30:10 -0500 (0:00:00.092) 0:01:35.259 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_name": "nopull", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:38 Saturday 11 January 2025 11:30:10 -0500 (0:00:00.057) 0:01:35.316 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 11 January 2025 11:30:10 -0500 (0:00:00.106) 0:01:35.422 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 11 January 2025 11:30:10 -0500 (0:00:00.077) 0:01:35.500 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 11 January 2025 11:30:10 -0500 (0:00:00.077) 0:01:35.577 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Saturday 11 January 2025 11:30:10 -0500 (0:00:00.151) 0:01:35.729 ****** ok: [managed-node2] => { "changed": false, "stat": { "atime": 1736612945.6352425, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1736612940.9742577, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 9160785, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "1643853349", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.416) 0:01:36.145 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.042) 0:01:36.188 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.040) 0:01:36.228 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.040) 0:01:36.269 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.041) 0:01:36.310 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.056) 0:01:36.367 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.053) 0:01:36.420 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.063) 0:01:36.484 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if no kube spec is given] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:43 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.066) 0:01:36.550 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_kube", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:52 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.062) 0:01:36.613 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:60 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.095) 0:01:36.708 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_path": "/etc/containers/ansible-kubernetes.d" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:64 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.140) 0:01:36.849 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_file": "/etc/containers/ansible-kubernetes.d/nopull.yml" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:68 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.061) 0:01:36.911 ****** ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Get service name using systemd-escape] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:75 Saturday 11 January 2025 11:30:11 -0500 (0:00:00.066) 0:01:36.977 ****** ok: [managed-node2] => { "changed": false, "cmd": [ "systemd-escape", "--template", "podman-kube@.service", "/etc/containers/ansible-kubernetes.d/nopull.yml" ], "delta": "0:00:00.005587", "end": "2025-01-11 11:30:12.313212", "rc": 0, "start": "2025-01-11 11:30:12.307625" } STDOUT: podman-kube@-etc-containers-ansible\x2dkubernetes.d-nopull.yml.service TASK [fedora.linux_system_roles.podman : Cleanup containers and services] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:83 Saturday 11 January 2025 11:30:12 -0500 (0:00:00.427) 0:01:37.404 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:2 Saturday 11 January 2025 11:30:12 -0500 (0:00:00.147) 0:01:37.552 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:10 Saturday 11 January 2025 11:30:12 -0500 (0:00:00.065) 0:01:37.618 ****** ok: [managed-node2] => { "changed": false, "enabled": false, "failed_when_result": false, "name": "podman-kube@-etc-containers-ansible\\x2dkubernetes.d-nopull.yml.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "basic.target \"system-podman\\\\x2dkube.slice\" systemd-journald.socket -.mount sysinit.target network-online.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BindLogSockets": "no", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanLiveMount": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DebugInvocation": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "A template for running K8s workloads via podman-kube-play", "DevicePolicy": "auto", "Documentation": "\"man:podman-kube-play(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698229248", "EffectiveMemoryMax": "3698229248", "EffectiveTasksMax": "22365", "Environment": "\"PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\\\x2dkubernetes.d-nopull.yml.service\"", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true /etc/containers/ansible-kubernetes.d/nopull.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true /etc/containers/ansible-kubernetes.d/nopull.yml ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/ansible-kubernetes.d/nopull.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/ansible-kubernetes.d/nopull.yml ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/podman-kube@.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "podman-kube@-etc-containers-ansible\\x2dkubernetes.d-nopull.yml.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13978", "LimitNPROCSoft": "13978", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13978", "LimitSIGPENDINGSoft": "13978", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LiveMountResult": "success", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureDurationUSec": "[not set]", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3233046528", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "[not set]", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "[not set]", "MemorySwapMax": "infinity", "MemorySwapPeak": "[not set]", "MemoryZSwapCurrent": "[not set]", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"podman-kube@-etc-containers-ansible\\\\x2dkubernetes.d-nopull.yml.service\"", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivatePIDs": "no", "PrivateTmp": "no", "PrivateTmpEx": "no", "PrivateUsers": "no", "PrivateUsersEx": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectControlGroupsEx": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "\"system-podman\\\\x2dkube.slice\" sysinit.target -.mount", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system-podman\\x2dkube.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22365", "TimeoutAbortUSec": "1min 10s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 10s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-online.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.podman : Check if kube file exists] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:28 Saturday 11 January 2025 11:30:13 -0500 (0:00:00.628) 0:01:38.247 ****** ok: [managed-node2] => { "changed": false, "stat": { "atime": 1736612966.5271692, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "d5dc917e3cae36de03aa971a17ac473f86fdf934", "ctime": 1736612965.874172, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 67109058, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1736612965.605173, "nlink": 1, "path": "/etc/containers/ansible-kubernetes.d/nopull.yml", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 217, "uid": 0, "version": "3709144835", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Remove pod/containers] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:38 Saturday 11 January 2025 11:30:13 -0500 (0:00:00.435) 0:01:38.682 ****** changed: [managed-node2] => { "actions": [ "/usr/bin/podman kube play --down /etc/containers/ansible-kubernetes.d/nopull.yml" ], "changed": true, "failed_when_result": false } STDOUT: Pods stopped: 8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844 Pods removed: 8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844 Secrets removed: Volumes removed: TASK [fedora.linux_system_roles.podman : Remove kubernetes yaml file] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:56 Saturday 11 January 2025 11:30:14 -0500 (0:00:00.649) 0:01:39.331 ****** changed: [managed-node2] => { "changed": true, "path": "/etc/containers/ansible-kubernetes.d/nopull.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:61 Saturday 11 January 2025 11:30:14 -0500 (0:00:00.445) 0:01:39.777 ****** changed: [managed-node2] => { "changed": true, "cmd": [ "podman", "image", "prune", "-f" ], "delta": "0:00:00.029819", "end": "2025-01-11 11:30:15.158062", "rc": 0, "start": "2025-01-11 11:30:15.128243" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:70 Saturday 11 January 2025 11:30:15 -0500 (0:00:00.442) 0:01:40.219 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 11 January 2025 11:30:15 -0500 (0:00:00.112) 0:01:40.332 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 11 January 2025 11:30:15 -0500 (0:00:00.065) 0:01:40.398 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 11 January 2025 11:30:15 -0500 (0:00:00.065) 0:01:40.464 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update containers and services] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:87 Saturday 11 January 2025 11:30:15 -0500 (0:00:00.064) 0:01:40.529 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:189 Saturday 11 January 2025 11:30:15 -0500 (0:00:00.061) 0:01:40.590 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:196 Saturday 11 January 2025 11:30:15 -0500 (0:00:00.058) 0:01:40.648 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle credential files - absent] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:202 Saturday 11 January 2025 11:30:15 -0500 (0:00:00.059) 0:01:40.708 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:211 Saturday 11 January 2025 11:30:15 -0500 (0:00:00.082) 0:01:40.790 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Saturday 11 January 2025 11:30:15 -0500 (0:00:00.112) 0:01:40.903 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Saturday 11 January 2025 11:30:16 -0500 (0:00:00.203) 0:01:41.106 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Saturday 11 January 2025 11:30:16 -0500 (0:00:00.108) 0:01:41.215 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Saturday 11 January 2025 11:30:16 -0500 (0:00:00.048) 0:01:41.264 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Saturday 11 January 2025 11:30:16 -0500 (0:00:00.065) 0:01:41.329 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Saturday 11 January 2025 11:30:16 -0500 (0:00:00.073) 0:01:41.403 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Saturday 11 January 2025 11:30:16 -0500 (0:00:00.071) 0:01:41.474 ****** ok: [managed-node2] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed-node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node2] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } ok: [managed-node2] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Saturday 11 January 2025 11:30:16 -0500 (0:00:00.145) 0:01:41.620 ****** ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Saturday 11 January 2025 11:30:17 -0500 (0:00:00.882) 0:01:42.502 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Saturday 11 January 2025 11:30:17 -0500 (0:00:00.066) 0:01:42.569 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages))", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Saturday 11 January 2025 11:30:17 -0500 (0:00:00.074) 0:01:42.643 ****** skipping: [managed-node2] => { "false_condition": "__podman_is_transactional | d(false)" } TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Saturday 11 January 2025 11:30:17 -0500 (0:00:00.134) 0:01:42.777 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Saturday 11 January 2025 11:30:17 -0500 (0:00:00.069) 0:01:42.847 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Saturday 11 January 2025 11:30:17 -0500 (0:00:00.066) 0:01:42.913 ****** ok: [managed-node2] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.023243", "end": "2025-01-11 11:30:18.279323", "rc": 0, "start": "2025-01-11 11:30:18.256080" } STDOUT: podman version 5.3.1 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Saturday 11 January 2025 11:30:18 -0500 (0:00:00.454) 0:01:43.368 ****** ok: [managed-node2] => { "ansible_facts": { "podman_version": "5.3.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Saturday 11 January 2025 11:30:18 -0500 (0:00:00.072) 0:01:43.441 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Saturday 11 January 2025 11:30:18 -0500 (0:00:00.127) 0:01:43.569 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "(podman_quadlet_specs | length > 0) or (podman_secrets | length > 0)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Saturday 11 January 2025 11:30:18 -0500 (0:00:00.068) 0:01:43.637 ****** META: end_host conditional evaluated to False, continuing execution for managed-node2 skipping: [managed-node2] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node2" } MSG: end_host conditional evaluated to false, continuing execution for managed-node2 TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Saturday 11 January 2025 11:30:18 -0500 (0:00:00.081) 0:01:43.719 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__has_type_pod or __has_pod_file_ext or __has_pod_file_src_ext or __has_pod_template_src_ext or __has_pod_template_src_ext_j2", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:96 Saturday 11 January 2025 11:30:18 -0500 (0:00:00.084) 0:01:43.804 ****** META: end_host conditional evaluated to False, continuing execution for managed-node2 skipping: [managed-node2] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node2" } MSG: end_host conditional evaluated to false, continuing execution for managed-node2 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:109 Saturday 11 January 2025 11:30:18 -0500 (0:00:00.096) 0:01:43.900 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 11 January 2025 11:30:19 -0500 (0:00:00.191) 0:01:44.092 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 11 January 2025 11:30:19 -0500 (0:00:00.073) 0:01:44.165 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 11 January 2025 11:30:19 -0500 (0:00:00.070) 0:01:44.236 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Saturday 11 January 2025 11:30:19 -0500 (0:00:00.090) 0:01:44.327 ****** ok: [managed-node2] => { "changed": false, "stat": { "atime": 1736612945.6352425, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1736612940.9742577, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 9160785, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "1643853349", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Saturday 11 January 2025 11:30:19 -0500 (0:00:00.448) 0:01:44.775 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Saturday 11 January 2025 11:30:19 -0500 (0:00:00.070) 0:01:44.845 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Saturday 11 January 2025 11:30:19 -0500 (0:00:00.069) 0:01:44.915 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Saturday 11 January 2025 11:30:19 -0500 (0:00:00.066) 0:01:44.982 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.066) 0:01:45.048 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.065) 0:01:45.114 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.053) 0:01:45.167 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.120) 0:01:45.287 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:115 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.042) 0:01:45.329 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:124 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.050) 0:01:45.380 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.072) 0:01:45.453 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.049) 0:01:45.503 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:127 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.065) 0:01:45.569 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.106) 0:01:45.675 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.055) 0:01:45.730 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:130 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.039) 0:01:45.770 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.129) 0:01:45.899 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Saturday 11 January 2025 11:30:20 -0500 (0:00:00.049) 0:01:45.949 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:133 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.068) 0:01:46.018 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.110) 0:01:46.128 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.066) 0:01:46.194 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.068) 0:01:46.262 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.066) 0:01:46.329 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:139 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.065) 0:01:46.394 ****** included: fedora.linux_system_roles.firewall for managed-node2 TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.162) 0:01:46.557 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed-node2 TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.124) 0:01:46.681 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.072) 0:01:46.754 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.042) 0:01:46.797 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.044) 0:01:46.841 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.043) 0:01:46.885 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 Saturday 11 January 2025 11:30:21 -0500 (0:00:00.043) 0:01:46.928 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: firewalld TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43 Saturday 11 January 2025 11:30:22 -0500 (0:00:00.748) 0:01:47.676 ****** skipping: [managed-node2] => { "false_condition": "__firewall_is_transactional | d(false)" } TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48 Saturday 11 January 2025 11:30:22 -0500 (0:00:00.044) 0:01:47.721 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53 Saturday 11 January 2025 11:30:22 -0500 (0:00:00.053) 0:01:47.774 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Saturday 11 January 2025 11:30:22 -0500 (0:00:00.055) 0:01:47.829 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Saturday 11 January 2025 11:30:22 -0500 (0:00:00.040) 0:01:47.869 ****** skipping: [managed-node2] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Saturday 11 January 2025 11:30:23 -0500 (0:00:00.144) 0:01:48.013 ****** ok: [managed-node2] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sat 2025-01-11 11:29:10 EST", "ActiveEnterTimestampMonotonic": "271605433", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus-broker.service dbus.socket sysinit.target polkit.service system.slice basic.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sat 2025-01-11 11:29:10 EST", "AssertTimestampMonotonic": "271333792", "Before": "multi-user.target shutdown.target network-pre.target", "BindLogSockets": "no", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "437791000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanLiveMount": "yes", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2025-01-11 11:29:10 EST", "ConditionTimestampMonotonic": "271333789", "ConfigurationDirectoryMode": "0755", "Conflicts": "iptables.service ipset.service ip6tables.service ebtables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4516", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DebugInvocation": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698229248", "EffectiveMemoryMax": "3698229248", "EffectiveTasksMax": "22365", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainHandoffTimestampMonotonic": "271363582", "ExecMainPID": "10862", "ExecMainStartTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainStartTimestampMonotonic": "271336631", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2025-01-11 11:29:10 EST", "InactiveExitTimestampMonotonic": "271337679", "InvocationID": "f2f268445be048bc90d6f558b31f85a8", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13978", "LimitNPROCSoft": "13978", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13978", "LimitSIGPENDINGSoft": "13978", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LiveMountResult": "success", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "10862", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureDurationUSec": "[not set]", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3210842112", "MemoryCurrent": "33013760", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "35106816", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "0", "MemorySwapMax": "infinity", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivatePIDs": "no", "PrivateTmp": "no", "PrivateTmpEx": "no", "PrivateUsers": "no", "PrivateUsersEx": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectControlGroupsEx": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket sysinit.target system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestamp": "Sat 2025-01-11 11:29:10 EST", "StateChangeTimestampMonotonic": "271605433", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22365", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Saturday 11 January 2025 11:30:23 -0500 (0:00:00.581) 0:01:48.595 ****** ok: [managed-node2] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sat 2025-01-11 11:29:10 EST", "ActiveEnterTimestampMonotonic": "271605433", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus-broker.service dbus.socket sysinit.target polkit.service system.slice basic.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sat 2025-01-11 11:29:10 EST", "AssertTimestampMonotonic": "271333792", "Before": "multi-user.target shutdown.target network-pre.target", "BindLogSockets": "no", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "437791000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanLiveMount": "yes", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2025-01-11 11:29:10 EST", "ConditionTimestampMonotonic": "271333789", "ConfigurationDirectoryMode": "0755", "Conflicts": "iptables.service ipset.service ip6tables.service ebtables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4516", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DebugInvocation": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698229248", "EffectiveMemoryMax": "3698229248", "EffectiveTasksMax": "22365", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainHandoffTimestampMonotonic": "271363582", "ExecMainPID": "10862", "ExecMainStartTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainStartTimestampMonotonic": "271336631", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2025-01-11 11:29:10 EST", "InactiveExitTimestampMonotonic": "271337679", "InvocationID": "f2f268445be048bc90d6f558b31f85a8", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13978", "LimitNPROCSoft": "13978", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13978", "LimitSIGPENDINGSoft": "13978", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LiveMountResult": "success", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "10862", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureDurationUSec": "[not set]", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3210100736", "MemoryCurrent": "33013760", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "35106816", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "0", "MemorySwapMax": "infinity", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivatePIDs": "no", "PrivateTmp": "no", "PrivateTmpEx": "no", "PrivateUsers": "no", "PrivateUsersEx": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectControlGroupsEx": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket sysinit.target system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestamp": "Sat 2025-01-11 11:29:10 EST", "StateChangeTimestampMonotonic": "271605433", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22365", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Saturday 11 January 2025 11:30:24 -0500 (0:00:00.545) 0:01:49.140 ****** ok: [managed-node2] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3.12", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Saturday 11 January 2025 11:30:24 -0500 (0:00:00.052) 0:01:49.192 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Saturday 11 January 2025 11:30:24 -0500 (0:00:00.042) 0:01:49.235 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Saturday 11 January 2025 11:30:24 -0500 (0:00:00.062) 0:01:49.297 ****** ok: [managed-node2] => (item={'port': '15001-15003/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "15001-15003/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Saturday 11 January 2025 11:30:24 -0500 (0:00:00.591) 0:01:49.889 ****** skipping: [managed-node2] => (item={'port': '15001-15003/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "'detailed' in fw[0]", "item": { "port": "15001-15003/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed-node2] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Saturday 11 January 2025 11:30:25 -0500 (0:00:00.105) 0:01:49.995 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'detailed' in fw[0]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Saturday 11 January 2025 11:30:25 -0500 (0:00:00.083) 0:01:50.078 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Saturday 11 January 2025 11:30:25 -0500 (0:00:00.069) 0:01:50.147 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Saturday 11 January 2025 11:30:25 -0500 (0:00:00.068) 0:01:50.215 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Saturday 11 January 2025 11:30:25 -0500 (0:00:00.063) 0:01:50.278 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Saturday 11 January 2025 11:30:25 -0500 (0:00:00.151) 0:01:50.430 ****** skipping: [managed-node2] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:146 Saturday 11 January 2025 11:30:25 -0500 (0:00:00.112) 0:01:50.542 ****** redirecting (type: modules) ansible.builtin.selinux to ansible.posix.selinux redirecting (type: modules) ansible.builtin.selinux to ansible.posix.selinux redirecting (type: modules) ansible.builtin.seboolean to ansible.posix.seboolean included: fedora.linux_system_roles.selinux for managed-node2 TASK [fedora.linux_system_roles.selinux : Set ansible_facts required by role and install packages] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:2 Saturday 11 January 2025 11:30:25 -0500 (0:00:00.253) 0:01:50.796 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml for managed-node2 TASK [fedora.linux_system_roles.selinux : Ensure ansible_facts used by role] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml:2 Saturday 11 January 2025 11:30:25 -0500 (0:00:00.126) 0:01:50.922 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Ensure SELinux packages] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml:7 Saturday 11 January 2025 11:30:26 -0500 (0:00:00.103) 0:01:51.026 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml for managed-node2 TASK [fedora.linux_system_roles.selinux : Check if system is ostree] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:5 Saturday 11 January 2025 11:30:26 -0500 (0:00:00.139) 0:01:51.166 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:10 Saturday 11 January 2025 11:30:26 -0500 (0:00:00.069) 0:01:51.235 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:17 Saturday 11 January 2025 11:30:26 -0500 (0:00:00.075) 0:01:51.311 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:22 Saturday 11 January 2025 11:30:26 -0500 (0:00:00.132) 0:01:51.443 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux python2 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:26 Saturday 11 January 2025 11:30:26 -0500 (0:00:00.071) 0:01:51.514 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_python_version is version('3', '<')", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux python3 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:35 Saturday 11 January 2025 11:30:26 -0500 (0:00:00.070) 0:01:51.585 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: python3-libselinux python3-policycoreutils TASK [fedora.linux_system_roles.selinux : Install SELinux python3 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:46 Saturday 11 January 2025 11:30:27 -0500 (0:00:00.793) 0:01:52.379 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_os_family == \"Suse\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux tool semanage] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:58 Saturday 11 January 2025 11:30:27 -0500 (0:00:00.047) 0:01:52.426 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: policycoreutils-python-utils TASK [fedora.linux_system_roles.selinux : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:72 Saturday 11 January 2025 11:30:28 -0500 (0:00:00.766) 0:01:53.193 ****** skipping: [managed-node2] => { "false_condition": "__selinux_is_transactional | d(false)" } TASK [fedora.linux_system_roles.selinux : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:77 Saturday 11 January 2025 11:30:28 -0500 (0:00:00.079) 0:01:53.272 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:82 Saturday 11 January 2025 11:30:28 -0500 (0:00:00.078) 0:01:53.351 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Refresh facts] *********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:89 Saturday 11 January 2025 11:30:28 -0500 (0:00:00.064) 0:01:53.415 ****** ok: [managed-node2] TASK [fedora.linux_system_roles.selinux : Set permanent SELinux state if enabled] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:5 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.922) 0:01:54.338 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_selinux.status == \"enabled\" and (selinux_state or selinux_policy)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set permanent SELinux state if disabled] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:13 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.036) 0:01:54.375 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_selinux.status == \"disabled\" and selinux_state", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set selinux_reboot_required] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:21 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.035) 0:01:54.410 ****** ok: [managed-node2] => { "ansible_facts": { "selinux_reboot_required": false }, "changed": false } TASK [fedora.linux_system_roles.selinux : Fail if reboot is required] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:25 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.100) 0:01:54.510 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_reboot_required", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Warn if SELinux is disabled] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:30 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.037) 0:01:54.548 ****** skipping: [managed-node2] => { "false_condition": "ansible_selinux.status == \"disabled\"" } TASK [fedora.linux_system_roles.selinux : Drop all local modifications] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:35 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.036) 0:01:54.585 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_all_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux boolean local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:43 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.037) 0:01:54.622 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_booleans_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux file context local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:48 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.035) 0:01:54.658 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_fcontexts_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux port local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:53 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.039) 0:01:54.698 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_ports_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux login local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:58 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.038) 0:01:54.737 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_logins_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set SELinux booleans] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:63 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.037) 0:01:54.774 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Set SELinux file contexts] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:74 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.036) 0:01:54.810 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Set an SELinux label on a port] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:87 Saturday 11 January 2025 11:30:29 -0500 (0:00:00.033) 0:01:54.844 ****** ok: [managed-node2] => (item={'ports': '15001-15003', 'setype': 'http_port_t'}) => { "__selinux_item": { "ports": "15001-15003", "setype": "http_port_t" }, "ansible_loop_var": "__selinux_item", "changed": false, "ports": [ "15001-15003" ], "proto": "tcp", "setype": "http_port_t", "state": "present" } TASK [fedora.linux_system_roles.selinux : Set linux user to SELinux user mapping] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:99 Saturday 11 January 2025 11:30:30 -0500 (0:00:00.620) 0:01:55.464 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Get SELinux modules facts] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:112 Saturday 11 January 2025 11:30:30 -0500 (0:00:00.043) 0:01:55.508 ****** ok: [managed-node2] => { "ansible_facts": { "selinux_checksums": true, "selinux_installed_modules": { "abrt": { "100": { "checksum": "sha256:5fdd1acb3c17ce67f51edc9160653adada9aade378218aed1b38acc0b4d4157a", "enabled": 1 } }, "accountsd": { "100": { "checksum": "sha256:50197ed40bbac67de44c12d745bc4a1259c8ae3f3828f5b4cfcd02b187ee1f47", "enabled": 1 } }, "acct": { "100": { "checksum": "sha256:14149edb1593dad6e684a672c57ca35f2ff51aee24ef503171c28671dc1bff13", "enabled": 1 } }, "afs": { "100": { "checksum": "sha256:769aaf25a8699233ceeda87ac858ebc91f094ea98b303c54e2c810760acd9756", "enabled": 1 } }, "afterburn": { "100": { "checksum": "sha256:9b9c00f12dea94989d2f5c592fa8ba4cff49b1ae7c9636fc38baa7957750e80f", "enabled": 1 } }, "aiccu": { "100": { "checksum": "sha256:d7ed1435298abcb42678af33c66ba2bac4c02723601c5018ecef64ca4c4e8f0b", "enabled": 1 } }, "aide": { "100": { "checksum": "sha256:95c6046b0d690df4792fe4418de5f7753279b72c78c5f1dab5397aba8bdf7152", "enabled": 1 } }, "alsa": { "100": { "checksum": "sha256:c0be471cd7182d1126cd6d19b276742a6ace62429425a81d42d8af4c24bcdbb4", "enabled": 1 } }, "amanda": { "100": { "checksum": "sha256:2ea0f86264cad0fa7f1eba0e6238526e92fadc65fd9754f62bc6d59b2da9086c", "enabled": 1 } }, "amtu": { "100": { "checksum": "sha256:1c77cc13007a0024e3121f064aad7dafce5fd57d7f71140f0af6b5af10805a8b", "enabled": 1 } }, "anaconda": { "100": { "checksum": "sha256:79386dbc8f95842f819ddbb01ce5824752d78bb4992783288d79db391a4ec3b5", "enabled": 1 } }, "antivirus": { "100": { "checksum": "sha256:48aafc3e308b4db80797ec75f52d3cc1b8216315d28108a10a03e47a0209adda", "enabled": 1 } }, "apache": { "100": { "checksum": "sha256:30a1e157e56f57a82557524173635e911f796fa8733f98ad190fc854aba893d8", "enabled": 1 } }, "apcupsd": { "100": { "checksum": "sha256:46a2e84d4dd0064f0b368787ed9c494ace4e337d5d80bcc6db77f66dede34375", "enabled": 1 } }, "apm": { "100": { "checksum": "sha256:4f5ff9b6ad61bc6a311bc71768e7ead7fe273989b7cf2fa58b1a13221e8d82f4", "enabled": 1 } }, "application": { "100": { "checksum": "sha256:ee11bb52bc37177920d348c4b920738a1b64413f516c66e18fa5a453a6d81f5e", "enabled": 1 } }, "arpwatch": { "100": { "checksum": "sha256:11a3c765598ae81cda7af561ada4795fb53a9ae9cfcb02af806b08cfc7f00b0e", "enabled": 1 } }, "asterisk": { "100": { "checksum": "sha256:ce887eb790f85810eb5eec0ac96a4e8028d633cf16fc117db4f733e6040c11c2", "enabled": 1 } }, "auditadm": { "100": { "checksum": "sha256:201655fdd63b544aa448c7b46c8b7ec8aa726ef3b55e9ef928ff3502c128973b", "enabled": 1 } }, "authlogin": { "100": { "checksum": "sha256:f1937fa4d112f634f69b9f0104872ce507c56928951a961c26ddf385a12be54e", "enabled": 1 } }, "automount": { "100": { "checksum": "sha256:5d308d3654fc5b0cea77156c27b98bbe2e370b843a920d75e1d616472caee55f", "enabled": 1 } }, "avahi": { "100": { "checksum": "sha256:ab1c62da7d38c1870efff54480b9a43ac2683eb0ebd77925a6fdea0fd47e207c", "enabled": 1 } }, "awstats": { "100": { "checksum": "sha256:f4a83008d582e61321a1d5462807c402d6cd3034d272dab00dd788122f245114", "enabled": 1 } }, "bacula": { "100": { "checksum": "sha256:04e3c063e146bd0f96587e9f49cdb4020e83bf69ba77250d61d2c9fcea53476b", "enabled": 1 } }, "base": { "100": { "checksum": "sha256:106220779e72010bf08e45ebec252a2e47257ca1134c07dbcd68a0d37e790931", "enabled": 1 } }, "bcfg2": { "100": { "checksum": "sha256:860a81a71f68a66e7dc0cee62e02f5da8d2dbb83724bfde9abc24b56b06ec914", "enabled": 1 } }, "bind": { "100": { "checksum": "sha256:ab2767aacd75df866fc6080d7db8dfefbd013a4ad578412ccfa6dece0d0b8ea7", "enabled": 1 } }, "bitlbee": { "100": { "checksum": "sha256:6ae2a654612810131bc874411156b1ef5d784143484e2a867ef413b6ab0f8750", "enabled": 1 } }, "blkmapd": { "100": { "checksum": "sha256:a6ac037c876fcea13d9457717138ec1c17913a0a303469219aa610351225c30c", "enabled": 1 } }, "blueman": { "100": { "checksum": "sha256:10372514dd684076a8cc4c93211d6f6732dcff0fa6aceed02dce868b12aaee82", "enabled": 1 } }, "bluetooth": { "100": { "checksum": "sha256:d8e9f771f666c02cacd7881f83cebd64ca88d79382fddca9e608d75df65099a2", "enabled": 1 } }, "boinc": { "100": { "checksum": "sha256:4ede0a1fb4501325ad8d0ff576f2f20bb0c3e994d2142372449de840f10e8b4b", "enabled": 1 } }, "boltd": { "100": { "checksum": "sha256:63a74e88f61411cfd03aa6e719867d1ae3238b99f3adb2f91cc0973e8b10e174", "enabled": 1 } }, "boothd": { "100": { "checksum": "sha256:4df95e4f92b1da81af245d74fd4481aad048554cad1d31f87985e3c68814ac8e", "enabled": 1 } }, "bootloader": { "100": { "checksum": "sha256:c4b351c2ad386a4c22a477788aeba9ae713cddaae623d733f1957d5d521c717d", "enabled": 1 } }, "bootupd": { "100": { "checksum": "sha256:fd688c68b40c042dd04fe9cb65c7a270353a2f7c677a5e73864b34ee6118f94d", "enabled": 1 } }, "brctl": { "100": { "checksum": "sha256:3a931c1c282cee1b53f6aee3730fef0e498fba3c5005218b4c7a923a7ce226c2", "enabled": 1 } }, "brltty": { "100": { "checksum": "sha256:0471f4ad4a09025a1a23ad398d869d88b90172a869800cf06660c886033c7409", "enabled": 1 } }, "bugzilla": { "100": { "checksum": "sha256:480fe3690a301d2359b747a03b7f65fd881063f6f7d877c55a16e6a51f324df6", "enabled": 1 } }, "cachefilesd": { "100": { "checksum": "sha256:04eafc0b0c7783758bba1cb55893416c204933bf0fdef964ce23c6f17c626c65", "enabled": 1 } }, "calamaris": { "100": { "checksum": "sha256:3b369138db2945167e6b10503d99902411fd5213b91dae978afe344b174240f7", "enabled": 1 } }, "callweaver": { "100": { "checksum": "sha256:0d66c1686efea20f479f6dd3057f5bd5550709d34e3a30745900abf363728bf1", "enabled": 1 } }, "canna": { "100": { "checksum": "sha256:5677c14448ec90dd564b7fe2c19146faa264a6cc80da75dba8f51c1e17af73d2", "enabled": 1 } }, "ccs": { "100": { "checksum": "sha256:ab0abd34c688230d836833a4f46228f3873ccbcfd3fd351d1306c203368ab63b", "enabled": 1 } }, "cdrecord": { "100": { "checksum": "sha256:c258276b231b731865f4237c8477010ea00b71e0e51501d7a206344381f625b1", "enabled": 1 } }, "certmaster": { "100": { "checksum": "sha256:5d46915b4fa32deccdf587b5970cde875d7b43643c09ec42a156193f9592c8ab", "enabled": 1 } }, "certmonger": { "100": { "checksum": "sha256:8746abfbb345c4bfd16b1c27e127a281e72388536aa9c7545ed308a266aceeb5", "enabled": 1 } }, "certwatch": { "100": { "checksum": "sha256:8d8299ad27c884a1657ff8ba52413f24867605e5e31c55a70659b3b0db4309d5", "enabled": 1 } }, "cfengine": { "100": { "checksum": "sha256:c3e0da50f4e8cd55d3a139afdcfa0fc5479b0e4afe78df4a1eab9400a71b64e8", "enabled": 1 } }, "cgroup": { "100": { "checksum": "sha256:e681dc6218ef1387435a6c4e95ed13e45d5edf2089991beb9bc8df41021cd865", "enabled": 1 } }, "chrome": { "100": { "checksum": "sha256:d64d5c77ef6d19be3aecf839fc444fab44a5717fa879c6d1bf6eb7564f556835", "enabled": 1 } }, "chronyd": { "100": { "checksum": "sha256:ea2b23dec628aecd68d6b8bc7d88d662e6bc0e72330f2a87e436f66a38b28ed4", "enabled": 1 } }, "cifsutils": { "100": { "checksum": "sha256:db66a0e17ecf7229eff4ffe84885bbff334085e3e20c282dd7e1eab97545e5e7", "enabled": 1 } }, "cinder": { "100": { "checksum": "sha256:95351f4464622aed05297db68b883d47165ce7db322134048f5a99479c734da3", "enabled": 1 } }, "cipe": { "100": { "checksum": "sha256:25016148cbfa7480cee4b614572bf7a364f6e45a5aca844bb9a87bba217b93a5", "enabled": 1 } }, "clock": { "100": { "checksum": "sha256:5f8b31db8708926fe3de99aadf34b5114184bff20956c7203c6329433bc24dc7", "enabled": 1 } }, "clogd": { "100": { "checksum": "sha256:b6bdd15aeae4574bffcd8a3c862261eb9e21ae5451035cc470ee9eba6cf49cd4", "enabled": 1 } }, "cloudform": { "100": { "checksum": "sha256:8d891dca8241582619e85b4d6f5aa273bd87d3bb7ff04199a38c91f47df12403", "enabled": 1 } }, "cmirrord": { "100": { "checksum": "sha256:e9ccb95f73f750fa464aa4fb5b76e5bad42050e27a128795826272ace5f70760", "enabled": 1 } }, "cobbler": { "100": { "checksum": "sha256:e090b7686819373bcad8a0cdfcf17dd453f7873dfe757742b4bee5c2be7f58d5", "enabled": 1 } }, "collectd": { "100": { "checksum": "sha256:96c655a73312da543bc2aa652b14079cb6fdac677618325d9f1ad83996dbeee1", "enabled": 1 } }, "colord": { "100": { "checksum": "sha256:180a83a6a6059ef9044f9a171ec71a5934a5a19da0f9ce5451c3ba1bbfb75343", "enabled": 1 } }, "comsat": { "100": { "checksum": "sha256:d1e10fcd329ec130a7346374dd6cf339eb38e2c37bc05f371c0381cc7e022d2e", "enabled": 1 } }, "condor": { "100": { "checksum": "sha256:f27781ec40ba6625c3854cf0e6c311a345d5619a3b07c7105635dc60a9369a6c", "enabled": 1 } }, "conman": { "100": { "checksum": "sha256:7984cbaf5572913966bae94db45941e7e15fc195f70c0bd30cbd6b8a24998687", "enabled": 1 } }, "conntrackd": { "100": { "checksum": "sha256:0607c19ac5763806c4420869caf11320ecad25b9ba9e10307105c75e9a26f6be", "enabled": 1 } }, "consolekit": { "100": { "checksum": "sha256:e6e39c814a93fa8973beffb8f891030f7b8094d9987e221d140c1a1fc79a196b", "enabled": 1 } }, "coreos_installer": { "100": { "checksum": "sha256:0ca9390a1bd0c5caa0958c03c1f665a129836f50e3ee11d54fa21800fab9aa08", "enabled": 1 } }, "couchdb": { "100": { "checksum": "sha256:e7ffb1afd19042410ffc4af01ee0fe31ff9f434d23451dd3bd01a7052be716b2", "enabled": 1 } }, "courier": { "100": { "checksum": "sha256:113abb8c428ab21653d9c2f9aef35354470381234ccb493e882ef22f72c2ede4", "enabled": 1 } }, "cpucontrol": { "100": { "checksum": "sha256:36939ecc2515a7980f3be7a2d9518018b3502aa526a446b04552f43a671518dc", "enabled": 1 } }, "cpufreqselector": { "100": { "checksum": "sha256:43b7f9d92ade2e28e26c7f9736d94d91b9a20812118f466b39aa4dc7a02e0959", "enabled": 1 } }, "cpuplug": { "100": { "checksum": "sha256:97a1c9de6df462cab0645b58e16072042c4cab7704c10209950684673a049b8d", "enabled": 1 } }, "cron": { "100": { "checksum": "sha256:0e02e661cc66e751e27d9ece16db633d5ab2036dab470b9dd4a1ef0725bf1534", "enabled": 1 } }, "ctdb": { "100": { "checksum": "sha256:7a7a3998f8a29178005386630556039bba9812f206b5f6782bdcfc0c22f66cbc", "enabled": 1 } }, "cups": { "100": { "checksum": "sha256:90605693f95fba47762dba603d99f0603ccea7238937bd0ed4ca8dc680fd7382", "enabled": 1 } }, "cvs": { "100": { "checksum": "sha256:05274fef587a5ac9def4bce8be2540bc893dd2c7b0052ad9f4e5de0c6da55718", "enabled": 1 } }, "cyphesis": { "100": { "checksum": "sha256:635739376db164b36a935a3bb0dbe8b5644b0fca8d3791b0efde3da36372a4aa", "enabled": 1 } }, "cyrus": { "100": { "checksum": "sha256:26567d47104e4634ab34aa39e0a3ea5f4ed85670b6c66b63e86c47c809080344", "enabled": 1 } }, "daemontools": { "100": { "checksum": "sha256:429c5e1b65232dbfb74741ba1514315288e2704c8a105759c31b85e3742200e1", "enabled": 1 } }, "dbadm": { "100": { "checksum": "sha256:24d6e3dfb505a297328108979f63cbad1b2a0f241c2fb2e84f93dadc086f1273", "enabled": 1 } }, "dbskk": { "100": { "checksum": "sha256:0bf031897570904400ae864c9d12f8880047d1c0368ebb8edc7fd7232f2eecab", "enabled": 1 } }, "dbus": { "100": { "checksum": "sha256:c0b2fa3d02d462ddbf8269485861a3e6e873cd5f05a174a99b121095cbd31f97", "enabled": 1 } }, "dcc": { "100": { "checksum": "sha256:f237213b8df023a69261f26318e3572f28a2c03bccf8d37505dabeca64aa9fe4", "enabled": 1 } }, "ddclient": { "100": { "checksum": "sha256:4159c1afa1a0e1977328780489eb61c30f61823b389f4cd0d79f25a7b27d5857", "enabled": 1 } }, "denyhosts": { "100": { "checksum": "sha256:643ce1a1f7dc2df66e3a1ff57d3dbd9631ea7c9b9b83467d7e4e2f4dc701186b", "enabled": 1 } }, "devicekit": { "100": { "checksum": "sha256:845cf9ac75f705a72df5d2b643b810fe266919369f6840e461f9d3043db97bb9", "enabled": 1 } }, "dhcp": { "100": { "checksum": "sha256:57fe91a9ccded7037429132ca5c28c408881f296eb8e603041b61f82c673c7c6", "enabled": 1 } }, "dictd": { "100": { "checksum": "sha256:246922587bf734f34d26ddb793ab9c649460b3cea63acfca54dc8a989e948d97", "enabled": 1 } }, "dirsrv": { "100": { "checksum": "sha256:9c67bc04c2d69254c0421c864b6f309c4c213bf415759b93a11a21bafa55b7f2", "enabled": 1 } }, "distcc": { "100": { "checksum": "sha256:a8179c2e64684b994ca8802badba7dbe0df42564f400307eeca0274f442cc0c4", "enabled": 1 } }, "dmesg": { "100": { "checksum": "sha256:977de79bba759fc421b791cc5496b0479b34077259eb0b9aab7049cce9e3d537", "enabled": 1 } }, "dmidecode": { "100": { "checksum": "sha256:dbe77a5d3671052e5b1dff63016c24e6e576803c82edbf2a913147a24f893124", "enabled": 1 } }, "dnsmasq": { "100": { "checksum": "sha256:7ad2d3372989319aaef08ac68aff266ae10dd262728284d4531851e893fe9edf", "enabled": 1 } }, "dnssec": { "100": { "checksum": "sha256:70ff660118b3bb9c8e5107d566adf6c2ad0b4af94d6757852c68a00eb19f6e56", "enabled": 1 } }, "dovecot": { "100": { "checksum": "sha256:1057c6ca706e9266c2e147865fb58e634439b7ec3e6c1faba0185401e2d4e985", "enabled": 1 } }, "drbd": { "100": { "checksum": "sha256:225633f600a5c31fcb19a9b721677833e32b88e57ef8cfaa3fa88b25bfc87c87", "enabled": 1 } }, "dspam": { "100": { "checksum": "sha256:503e9b509a0f764faea4a2415311aa1b43bfd8b78bbcc8090f6bc81c200beea8", "enabled": 1 } }, "entropyd": { "100": { "checksum": "sha256:e2aa3a7bc12f543b79842c2504cb1b53b73c55c78a42e24fbead6ab2deb120a4", "enabled": 1 } }, "exim": { "100": { "checksum": "sha256:b097146234c292a604496276a05ae2cbb09bbfdbf0ac9d7948b42afddd953347", "enabled": 1 } }, "fail2ban": { "100": { "checksum": "sha256:c51270c9abffd1d173532bbce08b28480f8e261d099370268b5364c343726140", "enabled": 1 } }, "fcoe": { "100": { "checksum": "sha256:a8a2c86fcc37532b9bf516e070ff53c9e675cb5e9f154966dde6a9b5fc6d7bc5", "enabled": 1 } }, "fdo": { "100": { "checksum": "sha256:7ecdacdeaddd502ea66686e0eaa24fa19fef7202f0092a7174db2aa86eddae8e", "enabled": 1 } }, "fedoratp": { "100": { "checksum": "sha256:85ee3ad2c50bf2b0ba1fb4fe0c57850f9ffd3bb5c11d11a5b5e159ac65783c18", "enabled": 1 } }, "fetchmail": { "100": { "checksum": "sha256:56ef87427ba73e1981192f595dd2c169b322cbed91da8de818e995434e07c757", "enabled": 1 } }, "finger": { "100": { "checksum": "sha256:9aa41b846e607a383c12d0bb4a9695058f8c9b1ed4b26bb59a96855015d9ab35", "enabled": 1 } }, "firewalld": { "100": { "checksum": "sha256:9e85ce564ebec8cae461d093727c4e339a32b10ce3efa569b5c2d53c27baae59", "enabled": 1 } }, "firewallgui": { "100": { "checksum": "sha256:433e16150ab0650ca51df171c1fbe1406cfe15dfa077bade2e48a7d8514bb0ba", "enabled": 1 } }, "firstboot": { "100": { "checksum": "sha256:a16615a4aea8b12fce51f6d7715627da3c3f6c6099db8c3087e29c920d81f861", "enabled": 1 } }, "fprintd": { "100": { "checksum": "sha256:1cb012beb208aebe26b53919c54e8e902210feb7b26d52b05cbe760d273b0645", "enabled": 1 } }, "freeipmi": { "100": { "checksum": "sha256:f62cf10a14cb8de0f158fed8cc9400a8769be1ffb2f3dcdcd0e920de20d68007", "enabled": 1 } }, "freqset": { "100": { "checksum": "sha256:5c96b7146bde0acec28d73afc25b0a35fbfffe2ccd69529717eca6927dc558a7", "enabled": 1 } }, "fstools": { "100": { "checksum": "sha256:0fdc4babffbc8fa18d9445b74b14fc08b77fd6e371c96dc1f5283d6ff7c5f546", "enabled": 1 } }, "ftp": { "100": { "checksum": "sha256:dae3209382de68fdb49a72792108ec3a55c0e50abdfd38c5fae300da4c561c47", "enabled": 1 } }, "fwupd": { "100": { "checksum": "sha256:4f1c84d8ccd64b0af49edc9247ac92f49545fe4dfb24906acc133ebd7ba78bb9", "enabled": 1 } }, "games": { "100": { "checksum": "sha256:7d9b09309f694125d4ba8e1f5683baedfc2c935f0242be94531b43d842509ae9", "enabled": 1 } }, "gdomap": { "100": { "checksum": "sha256:bec0b003cf82cc2fe7e6f196bd82b929e87a827fe8d46bca5b1df2613a409d6b", "enabled": 1 } }, "geoclue": { "100": { "checksum": "sha256:3e8a90189684079f66e782c8e001c435ada771766dd1fa9ae33a977405f6ba99", "enabled": 1 } }, "getty": { "100": { "checksum": "sha256:e5ecf805b71dfe0f039f8eeefa730893c5417fe53b0e0eeba55f99670b4c4584", "enabled": 1 } }, "git": { "100": { "checksum": "sha256:69850d2342f0fcb96a719d39a3698d547c105880e6834f232a4d845a18a046f3", "enabled": 1 } }, "gitosis": { "100": { "checksum": "sha256:24935d3495375c7f6e0a0ddc091cf1f3f276ee155d07760bd7ba0c3ac115f6e8", "enabled": 1 } }, "glance": { "100": { "checksum": "sha256:fe00543de47f940ba5bf1ab84b7dc7f6a680053ef148329f2a9ec827eb47c61c", "enabled": 1 } }, "glusterd": { "100": { "checksum": "sha256:4d2befdcd924820622a0c6653772a9f5ba96aad9a3bd7b4d86daab96b233c9ea", "enabled": 1 } }, "gnome": { "100": { "checksum": "sha256:a5e2adeeda1a5b359025d956cbf20eb3773341207c72cdf7bf5001deeedf4712", "enabled": 1 } }, "gnome_remote_desktop": { "100": { "checksum": "sha256:d926eb26b34a6253b249bc1cea7696dae87887739ef77cd8cb92bdf7f6d1f371", "enabled": 1 } }, "gpg": { "100": { "checksum": "sha256:a9180cca4098d53514cd5f701f4afd40420a3664e8ffa86797ee3a8fadd9a84a", "enabled": 1 } }, "gpm": { "100": { "checksum": "sha256:2b5e83f7513a2470766cbaad46e28604ed8fcd6b640dc0350d90731a807f56ad", "enabled": 1 } }, "gpsd": { "100": { "checksum": "sha256:b64db8adeb9e8602211a0924c1b4b917d0f701df86984e73412344c9b3c2f5d9", "enabled": 1 } }, "gssproxy": { "100": { "checksum": "sha256:a43f376f2b1605cf0d45884b818db045eb37b4f037ae7af36ec5bf993676f4e5", "enabled": 1 } }, "guest": { "100": { "checksum": "sha256:262ccc7c3b237e7ebc31631d83baf2c6cbf8438cce02ce82f008fd6a3ff81a43", "enabled": 1 } }, "hddtemp": { "100": { "checksum": "sha256:fc86a156d59e31b02ce40cbda3c24b0c0baae2ed6f0cabf84cb9a3248ccd236e", "enabled": 1 } }, "hostapd": { "100": { "checksum": "sha256:ba13f04311a741f26ac34cd39688e2f4e75a546358c63c57a2d6a916dfc18315", "enabled": 1 } }, "hostname": { "100": { "checksum": "sha256:7366a48a1774b579953d8b8ae05f9ebe07e5d27984c995afe27509eacf1d79d6", "enabled": 1 } }, "hsqldb": { "100": { "checksum": "sha256:a0c31e8462c7fbbc6579270e5e35923deff56790f565157cb9c32c4eb3d6958b", "enabled": 1 } }, "hwloc": { "100": { "checksum": "sha256:fde041b834e5939945fc0d91e45940745254e128d96625ca5ab4d6e7a4379e27", "enabled": 1 } }, "hypervkvp": { "100": { "checksum": "sha256:520caa8d38f07a36d096f7ee8eb1b0610f49eafedca1721b340f5cdf21e23944", "enabled": 1 } }, "ibacm": { "100": { "checksum": "sha256:4bc0f9a40691f1abb15ce6cd277d7bab4b18720373a197c8310aa63995bcda0f", "enabled": 1 } }, "ica": { "100": { "checksum": "sha256:5daeb045f8b76fed05c1145d131f77dd556de131f777b7e259d8ec57787e7da6", "enabled": 1 } }, "icecast": { "100": { "checksum": "sha256:281ed7601c5b057802da5ffdc24a7e3308e0bb2591a263186f51e35816dffc12", "enabled": 1 } }, "iiosensorproxy": { "100": { "checksum": "sha256:05dd0967ab5bc2422ac57f2ec8a354181bf7038f0d1aabc2775c7c915bbbeaf3", "enabled": 1 } }, "inetd": { "100": { "checksum": "sha256:52f6df31e38b0c20b9fc264115ce56b4d6d8048dce5b6f250a6ca1b0675a0569", "enabled": 1 } }, "init": { "100": { "checksum": "sha256:7141bbe724e0c7f4b3dd7f335ab21267676b126f0443da9cbfa58f1372deb59c", "enabled": 1 } }, "inn": { "100": { "checksum": "sha256:78ae9628e756da216964c88e987dd91e40ec2803ff209763f628a434c8bf3b90", "enabled": 1 } }, "insights_client": { "100": { "checksum": "sha256:c6e547e12ca21f1f68ba74d667fab1bcbfc8d0461fc3a560e80160f32638ddba", "enabled": 1 } }, "iodine": { "100": { "checksum": "sha256:5c8ef0d8e730a3baca51f4c64b13df56149ef304effdc71fe0fb63f90e5a7296", "enabled": 1 } }, "iotop": { "100": { "checksum": "sha256:114de54fd32be9c32e88348ddb168291c0aaaba3a7937501f49e514585228092", "enabled": 1 } }, "ipmievd": { "100": { "checksum": "sha256:5777d2cc5e62942022a8e914847dc7436264585e41e2ce703c49c735e44add8c", "enabled": 1 } }, "ipsec": { "100": { "checksum": "sha256:ffde25d3dd0fa80b933b30cd457ea900389f84d13d4ea629f54c0be59136d4a6", "enabled": 1 } }, "iptables": { "100": { "checksum": "sha256:4c92a7102c209b0edf571a3ca3b35c597bd427d0d258c6829bfd29358448598b", "enabled": 1 } }, "irc": { "100": { "checksum": "sha256:291a822392f8a93d820aa03613bd0c058cfd1118fd6cca93a12175456eb2cf72", "enabled": 1 } }, "irqbalance": { "100": { "checksum": "sha256:2fc17d69e2cba360a0d9043585a3799f86697bf9189fcde62346b41bcf5bd9d9", "enabled": 1 } }, "iscsi": { "100": { "checksum": "sha256:a91d9d18f0502679892047e7a6023107b77412ae4b3d25f75982f766c792be58", "enabled": 1 } }, "isns": { "100": { "checksum": "sha256:5372fd0c432735b6ec28e0a5cafeebd74688912c7703616ca46b77287385c564", "enabled": 1 } }, "jabber": { "100": { "checksum": "sha256:d1916b00ece1c2b546cc04a7b6695f6459724a6bb7a92be78f399f61a613af86", "enabled": 1 } }, "jetty": { "100": { "checksum": "sha256:c78ba53bd379a65017e29895da717fa8e758494e94113b7f324d73121cb35423", "enabled": 1 } }, "jockey": { "100": { "checksum": "sha256:4cee27b59fe57c672587d220f3061417207a02bc0ff36188148f5c5f3101c1ad", "enabled": 1 } }, "journalctl": { "100": { "checksum": "sha256:27aeef17f22b9e891c4c793870f1124ec69ec7dc8f532bfec5b2a0e7319866cb", "enabled": 1 } }, "kafs": { "100": { "checksum": "sha256:602c3c40b28b653e11110e83c7f63122a75b9ad5cb7605299f6951a485284727", "enabled": 1 } }, "kdump": { "100": { "checksum": "sha256:bac33f3ca081f5f4b58fbbdbd3630aef7fb77c19368ec2cef66494cb8ef7fd90", "enabled": 1 } }, "kdumpgui": { "100": { "checksum": "sha256:7e16a86780b3b3d8a6afc5a09335cd6bf6bb39eec53815af9bbc53844655846c", "enabled": 1 } }, "keepalived": { "100": { "checksum": "sha256:5e76b1d8c27eea054956c86458f44ebc46d803035e2b3c0af67fd5feae26b9f2", "enabled": 1 } }, "kerberos": { "100": { "checksum": "sha256:65a81f3d94068c6e7cd3c1092cfbc080f853c7e02490356f7f16d40a3e8ace49", "enabled": 1 } }, "keyboardd": { "100": { "checksum": "sha256:e267709b6fc5f39679c3156b764ac82facb0b228e5b300d891f4a182662a1a5b", "enabled": 1 } }, "keystone": { "100": { "checksum": "sha256:1a7021842ad7350bc5674a455f4232b831ec8153d14635eba8930e8ade0980b7", "enabled": 1 } }, "keyutils": { "100": { "checksum": "sha256:0dc521ba68e3d21396cf5bce20427e6c0c5dfe635c743ccd4374beadd1a96cc5", "enabled": 1 } }, "kismet": { "100": { "checksum": "sha256:4398910d875745fbd61926116eddc0c7f3f6da3cf8b62fb4149d78734c344df3", "enabled": 1 } }, "kpatch": { "100": { "checksum": "sha256:36ae9280a2f7e86befeaf301f1d47087b9214bc97e79562105fcaa1246512467", "enabled": 1 } }, "ksmtuned": { "100": { "checksum": "sha256:66f6fee140f5f767906ea257b32ede5d214b8da0db4031313eb1fddb9a589d21", "enabled": 1 } }, "ktalk": { "100": { "checksum": "sha256:e783786109e8dac25e8be80e5782ec260343cf0db5b03126b85811a101bd9d0e", "enabled": 1 } }, "l2tp": { "100": { "checksum": "sha256:55f189e95389ac2f9205270f3dd53e25761d8e626989bdfb3c05b19c0086dc30", "enabled": 1 } }, "ldap": { "100": { "checksum": "sha256:cb97aa280c3bb13b6c8804e2a2bf971250db72b77ba5ece99702ab6434c3d14a", "enabled": 1 } }, "libraries": { "100": { "checksum": "sha256:bb45e05979c90858a0ed2ed0ecae78cc4e99b750145947a2940eb0493249d161", "enabled": 1 } }, "likewise": { "100": { "checksum": "sha256:f535c1e16dc83389fd5a2fa4716615c687a931e3a9ad7f65dc8607ef305f7994", "enabled": 1 } }, "linuxptp": { "100": { "checksum": "sha256:0d6f39eec4849cc1f5ca0014c29399b78317216ed0e73dfc524be3194ff1bc08", "enabled": 1 } }, "lircd": { "100": { "checksum": "sha256:cdd2ad3d9fa15b37c76e5bf0d781314f3d50897ef6cf377d70706d8a139639f7", "enabled": 1 } }, "livecd": { "100": { "checksum": "sha256:46a683c9c0693f57e0a687926c0ea32a8a176eaa168d1bf86e66463395b314cf", "enabled": 1 } }, "lldpad": { "100": { "checksum": "sha256:9fa26b1dd3ca8e52c93a1320fb66fceb0c8e83730227023d221a6a426b9dddf0", "enabled": 1 } }, "loadkeys": { "100": { "checksum": "sha256:8e876286cd84dd189d40234f691174bc5e0c214056ff0a9f454400cdafdcd6cd", "enabled": 1 } }, "locallogin": { "100": { "checksum": "sha256:f90a7ca564fcff2da3a9effa5c50e8c2a92de3e4c3831f741bc3f8cfc5be5808", "enabled": 1 } }, "lockdev": { "100": { "checksum": "sha256:d181bc52a0d2af9fe776462be417f9f28920a71f01f9894557e9bf7e128ddb5c", "enabled": 1 } }, "logadm": { "100": { "checksum": "sha256:c573637379d69beb37674b0ce9da5f57e60238a88ddad2faa1c3d456cd7df71a", "enabled": 1 } }, "logging": { "100": { "checksum": "sha256:d3494d1da872d6c128a1ab751a7ec84abb1ec73dccbfae200707935617fb8888", "enabled": 1 } }, "logrotate": { "100": { "checksum": "sha256:97dfef99f123dd0474194e78c4e8760df60d46db61f8460161bbcf4b78f7c530", "enabled": 1 } }, "logwatch": { "100": { "checksum": "sha256:c68409b41b76b32647af8f19d1f8f6ebd43430dc820e55d2b094730950a96419", "enabled": 1 } }, "lpd": { "100": { "checksum": "sha256:24594a8f8d0e864712707e8c4114f6d68dd19824045a9f3d10b8f590fc506858", "enabled": 1 } }, "lsm": { "100": { "checksum": "sha256:7df22b87a981a09999b90bb4ecea9713b595814a385fce77eb67e3c20042aa09", "enabled": 1 } }, "lttng-tools": { "100": { "checksum": "sha256:9f30b434e6c08d9a155bbc24ef6d71745d8d252c96d0ded0ccb6d158354c6546", "enabled": 1 } }, "lvm": { "100": { "checksum": "sha256:f0b2c006055d609858e59815fcd002fab6ce4a77761dfd2cc18b56219bf56fc1", "enabled": 1 } }, "mailman": { "100": { "checksum": "sha256:50901ef3e7ccc1828edc3fde37485af69afe28048ba25b7d9e3578b063d559d7", "enabled": 1 } }, "mailscanner": { "100": { "checksum": "sha256:914891424131853f76600695c57ec6283f9cc3c501c06e1427ef99f83431d1cd", "enabled": 1 } }, "man2html": { "100": { "checksum": "sha256:a7af8f4be03248a6d91c5df6c2becd22e3f45a71cecd671d44d2830fcd9f4a3b", "enabled": 1 } }, "mandb": { "100": { "checksum": "sha256:4afad481406029752a7960f97c253b41b8251018b219e16a858ccd029f9bb5f2", "enabled": 1 } }, "mcelog": { "100": { "checksum": "sha256:047b1501df33727c4257a655f1ad55aaed7da1d763782f7da45dfb1b19fa6a02", "enabled": 1 } }, "mediawiki": { "100": { "checksum": "sha256:7ce2990caacbf36b3ea6481def0630fa02e220ea9a4e2dc107251bce1368e569", "enabled": 1 } }, "memcached": { "100": { "checksum": "sha256:577518e3f4a348aaac570906781e3db46d9d53bf52d1a6f83c9f520dab8270e7", "enabled": 1 } }, "milter": { "100": { "checksum": "sha256:aa6812766cbdfad665abf5ad9633ffb945822d31714ffaa2bf4cb061f91e6b6c", "enabled": 1 } }, "minidlna": { "100": { "checksum": "sha256:f13c8b27b597e55d9e6342c756bc4d21de401fcbc4cb01e16c71c01a816c2e2b", "enabled": 1 } }, "minissdpd": { "100": { "checksum": "sha256:18649df6879516ea838eabbd19d6f7aefe2bb7da501c0b471c453d80ea4e02cb", "enabled": 1 } }, "miscfiles": { "100": { "checksum": "sha256:6f5792cde030e56b55f4b00cb53bd6d3724181c80aaaee5806e88c6a5298bb6c", "enabled": 1 } }, "mock": { "100": { "checksum": "sha256:3fdad88af11ebff86688d4db404640a1371af801e4a9b2f14a1b2abc1c76a503", "enabled": 1 } }, "modemmanager": { "100": { "checksum": "sha256:55370db1ae9a5e7ec0b4ec118f45ed61578656c6df3deb7b9e618362ddce4e29", "enabled": 1 } }, "modutils": { "100": { "checksum": "sha256:37b0e80e0d823ecd2606ece25c0b93eaea9f16da35261d79630b3e95c33a5fb4", "enabled": 1 } }, "mojomojo": { "100": { "checksum": "sha256:8ddaca2a54241adf5d42f9205cd073c7062e7692f31ebe15eb9e988e12efc9d4", "enabled": 1 } }, "mon_statd": { "100": { "checksum": "sha256:7e1bf20f24f949d86cf2cf1ecd59aef5aee94c9258d32a8fb4ee7646e4dc67e7", "enabled": 1 } }, "mongodb": { "100": { "checksum": "sha256:051c9eb43df88894edffecf5083ac5065a7051b743068b6f1866d923bca5ab95", "enabled": 1 } }, "motion": { "100": { "checksum": "sha256:154f6949eddb0f008a3b9166f736dde00bbc37df415f07c1263cd63ee609a5fd", "enabled": 1 } }, "mount": { "100": { "checksum": "sha256:acb9c37d19d0aec956cffe6f83d9593cb70bf8ad71ecf956e074b590286f489b", "enabled": 1 } }, "mozilla": { "100": { "checksum": "sha256:0a4982ad6ec48750689dc3c5d7ee946fca86c0c8b3763ee322aae33546109bed", "enabled": 1 } }, "mpd": { "100": { "checksum": "sha256:8807a444e6651ef4e67d0fed3bbc2558dc34b15e7293bfa03487d0a2c54ba905", "enabled": 1 } }, "mplayer": { "100": { "checksum": "sha256:3f999e549842f5a13f5f0d5aa4a752e74216c123f34f1797588b73d504afdc81", "enabled": 1 } }, "mptcpd": { "100": { "checksum": "sha256:f715709d773d8da3d99b4a7de877c7d7a8964d7a1342a61779edcdde2c75702d", "enabled": 1 } }, "mrtg": { "100": { "checksum": "sha256:740151127f1bcdb7699a6317eca571399e7cf58a755afd20dd141aec3035ab53", "enabled": 1 } }, "mta": { "100": { "checksum": "sha256:e04f6d7a528f9eeb7835428fe62d15043b9f4642cded56d682551e18920d04b3", "enabled": 1 } }, "munin": { "100": { "checksum": "sha256:1d2ea7b3403a2c521005ef980913298f01b2c00338366997273344590d0b218e", "enabled": 1 } }, "mysql": { "100": { "checksum": "sha256:07bc412265c70dcb7a3f85ba5265573ef76373364bbbd165ad3abb662e4de705", "enabled": 1 } }, "mythtv": { "100": { "checksum": "sha256:4cab8df79dc5a8e73ece06c418cb3ca58c74f669516b810de46b9f268d78ee8c", "enabled": 1 } }, "nagios": { "100": { "checksum": "sha256:712912fdf95fbf2adc1cb7427524d5536f0cd591bf3e77b63ef069213ae86704", "enabled": 1 } }, "namespace": { "100": { "checksum": "sha256:c032c4722dc3ec48bc9f196ae3de9ef9fd1c04554aa0913d99ade36ac0cf85af", "enabled": 1 } }, "ncftool": { "100": { "checksum": "sha256:e23fc9d15551fc1d7ba386350226d425282fcc090b9dfdcdb7065ebf3385666b", "enabled": 1 } }, "netlabel": { "100": { "checksum": "sha256:d9b7f3b8cb9b0ca564b43893085106dab4a6f6e7bee3143d778fc28218910435", "enabled": 1 } }, "netutils": { "100": { "checksum": "sha256:e828ed9b31086b0b53a5cf66cb2a580aae3e3290156b0d32c07eec5b1bd9cfe5", "enabled": 1 } }, "networkmanager": { "100": { "checksum": "sha256:bc61c1cee0f527dce7154fefb7906e6262e81d84108c7999f8db9cba7e4eb44c", "enabled": 1 } }, "ninfod": { "100": { "checksum": "sha256:4ed636e37871cc6a707bd29f6d990f4ff13d98bac2d8bf37922983d6289b3411", "enabled": 1 } }, "nis": { "100": { "checksum": "sha256:bce01b1750e64ca5f88e7d00d7d4dc88d5ebd430561f9b3cc680484b631c6c8a", "enabled": 1 } }, "nova": { "100": { "checksum": "sha256:16ebf87ef7e152147a0aa89e34161622d387348d4376f2eb91d971ef8ab476a0", "enabled": 1 } }, "nscd": { "100": { "checksum": "sha256:c7e5598679b39499dd00ac032b9a85b6e7e367bc64a0e2827cd799bb7e920a72", "enabled": 1 } }, "nsd": { "100": { "checksum": "sha256:75e019e0b2eda6179a3b39defbd43329dff4627e2c6a06c868ebee8b584332ee", "enabled": 1 } }, "nslcd": { "100": { "checksum": "sha256:69c6eaa2ef51784392d46e962c0f9b9439616dc97ba3105ff54e275a7b1e9db4", "enabled": 1 } }, "ntop": { "100": { "checksum": "sha256:cd4113e904b08b55f2a611ab02edbba284a91097a468b0d6b0d98a5dff397220", "enabled": 1 } }, "ntp": { "100": { "checksum": "sha256:5f188e0c5dabce04e39700cab0a3052cde82676b66e13b054d55f6f78f8d4b30", "enabled": 1 } }, "numad": { "100": { "checksum": "sha256:ed1193b412ae24e79c9a91531c99520b68ded7151bbf0f8cf0353d0dab713dec", "enabled": 1 } }, "nut": { "100": { "checksum": "sha256:54c04020d40b310f4b85f747395b3183859d816a6f181e4a54f5bd6f23b660df", "enabled": 1 } }, "nvme_stas": { "100": { "checksum": "sha256:1098f37f7b13bb6fa26d45bd4c54fd8cb38eddd1b4d02efaf88241f7338022c2", "enabled": 1 } }, "nx": { "100": { "checksum": "sha256:a8b41cc8424e4d9d67310c486221a9fc2701a9508104bdeb076f2ece6c94957c", "enabled": 1 } }, "obex": { "100": { "checksum": "sha256:4810ca21adc14e1f8b8f2a5b91e628038786e71ff84a2b3d701cdcafc6baf16c", "enabled": 1 } }, "oddjob": { "100": { "checksum": "sha256:d09199963de83937d060d3199ad48bf9f6753507c2748ad11d86628469f1e5b5", "enabled": 1 } }, "opafm": { "100": { "checksum": "sha256:020fbfc8bc4b762f26e66c399e2ae1f202c409d7ae1f10ef84f94c9672a1e55f", "enabled": 1 } }, "openct": { "100": { "checksum": "sha256:fd4ba492c3bc1b35ee1d9e4cf2b1fa21ac80ed31c9c76cb4503996f7c6367898", "enabled": 1 } }, "opendnssec": { "100": { "checksum": "sha256:cd5781c1d95be89112ad12c0d1341cc43b60259d4eea9b11442bc19da8ee462c", "enabled": 1 } }, "openfortivpn": { "100": { "checksum": "sha256:9a125b8b0328b3f01aea6aa395ce5117bc3a683e2b45b12e69dc6a95116ce203", "enabled": 1 } }, "openhpid": { "100": { "checksum": "sha256:65ec5e445fa5dd1471b83bccc319e9403a1c5888423442382751b1bf91c5188d", "enabled": 1 } }, "openshift": { "100": { "checksum": "sha256:07bea1192b1c51dd443db5a17d844ba3d7383abc3a3e7e979dcf0143753af074", "enabled": 1 } }, "openshift-origin": { "100": { "checksum": "sha256:88873289a3c1bd0c0c5d636b689bd8e1fe194ca528720c3191fb61b2ca024b50", "enabled": 1 } }, "opensm": { "100": { "checksum": "sha256:76a9f4d16f6b8fe2cb94b30b0c557dd4b067a4cb49502ff812b95290539062f2", "enabled": 1 } }, "openvpn": { "100": { "checksum": "sha256:ec0e41fdd7c04028cedf656a3fd60533adc778663a70734f5f2d569f5e79f53b", "enabled": 1 } }, "openvswitch": { "100": { "checksum": "sha256:1ab8ca3123bd0bbfbe2baac2e44124259401557c590cb8e906b06f28a4b9e32d", "enabled": 1 } }, "openwsman": { "100": { "checksum": "sha256:417b01a65501649535510d72e5fca7174183b5c873c1ddcb6921c052c8be3fc3", "enabled": 1 } }, "oracleasm": { "100": { "checksum": "sha256:6070bb82ad7b693e1ce70bcec6b975c4f5baec8a51e879f27e2c80ff2e076755", "enabled": 1 } }, "osad": { "100": { "checksum": "sha256:7bd10b072510204cc5f32e694b6eb1389fc20c40186fad86b76079e774023f4b", "enabled": 1 } }, "pads": { "100": { "checksum": "sha256:aeae1d20fad2ecc5e2f3ac84829becd50d70a7c41e8183c5b817e0d3f97f9484", "enabled": 1 } }, "passenger": { "100": { "checksum": "sha256:93cb4d1c0096f51325fa10c42ab72db38b5974c42b7d9fa6789482f56ce688c6", "enabled": 1 } }, "pcm": { "100": { "checksum": "sha256:4bb8d2c73a705f3051ce26155740950675bb556b11e2a7f59bb390ebbf7f9a87", "enabled": 1 } }, "pcmcia": { "100": { "checksum": "sha256:722b33d72298e2c8fa2900f1592c92673d098fcb6cd5802449086c918ca104dd", "enabled": 1 } }, "pcp": { "100": { "checksum": "sha256:a34e540529562d554136e3710fe67e8ca6e9ccee02b76dab369a448b58236f34", "enabled": 1 } }, "pcscd": { "100": { "checksum": "sha256:f50088e07d5fd52c3ad752e30852f0d3085cae9f13d2da6fd8c8141f1568ca28", "enabled": 1 } }, "pdns": { "100": { "checksum": "sha256:2d20aff28dd9cac707097c581edf455d754629927a32cc07854eb9baedb23b0a", "enabled": 1 } }, "pegasus": { "100": { "checksum": "sha256:81bedf6564a2a1e2b7ef03a9e5f646184f770202ad403e031f430e494abf8e88", "enabled": 1 } }, "permissivedomains": { "100": { "checksum": "sha256:2453bad4ace526f3cf2c60b358e95a5476692ef25da107b10f52f3af27c056d2", "enabled": 1 } }, "pesign": { "100": { "checksum": "sha256:544140731f1c09b504a6b85ec6ffbdda47c549c9858d6c2264498bfc56fd6127", "enabled": 1 } }, "pingd": { "100": { "checksum": "sha256:a4f1dd19ad8877fb1b48c431790bc350af44d3e2d0d08d59497d2a4425ce2d80", "enabled": 1 } }, "pkcs": { "100": { "checksum": "sha256:c0fe90911ac32abd43e43d9e25dd5a15c85301a1ad0ee579fd3997b9c25ad201", "enabled": 1 } }, "pki": { "100": { "checksum": "sha256:16249004873cf9214f949f73481b5476c0ab03e349542512af9b57aeb637c047", "enabled": 1 } }, "plymouthd": { "100": { "checksum": "sha256:4e33d3ebd3e3cd1fc439ca7bcf3c11086799c7af6a1e12460eb4fc5bab23b0be", "enabled": 1 } }, "podsleuth": { "100": { "checksum": "sha256:7f525e710a117aeaec78a9ab1cbbdd54a150ffe26ed5a1bc29595824fd9e1931", "enabled": 1 } }, "policykit": { "100": { "checksum": "sha256:fe85c9803d13edb28fd5a47803918af2a1e70d66b57e920fefe8e6cfdf778171", "enabled": 1 } }, "polipo": { "100": { "checksum": "sha256:8ab8b8f9438c6ae6c96323b6cf5937528d8223c38629f9bed4a2df00f4b9b686", "enabled": 1 } }, "portmap": { "100": { "checksum": "sha256:59c39477c07ba01841c150a2232e21074575ad3f93782e73b35da8f7acb121af", "enabled": 1 } }, "portreserve": { "100": { "checksum": "sha256:00aa972aa3ea3cf49a0134ebc75df0d6b7eb87ca7b31c61b3107001111abda8a", "enabled": 1 } }, "postfix": { "100": { "checksum": "sha256:68f178fd86c386a5d4df236f983239aa04eb6482304df3e9eb56340b744f5d8c", "enabled": 1 } }, "postgresql": { "100": { "checksum": "sha256:bab7ecb38d7aa40fc1de9d3f1a06e0f83c473e22e2f2391cbdc50e27ef92f7dd", "enabled": 1 } }, "postgrey": { "100": { "checksum": "sha256:8e1915326802644782f3041e65e3b1dfb6f534f2a66f4cb00ca8391f35ef2048", "enabled": 1 } }, "ppp": { "100": { "checksum": "sha256:912c8757c9d3cc46657aa135743c58836db26f4a6abf17807017c86a8512d90f", "enabled": 1 } }, "prelink": { "100": { "checksum": "sha256:0b0ad935c694badc8da9368226c53640cfade72c6828d094924029340ad5521e", "enabled": 1 } }, "prelude": { "100": { "checksum": "sha256:92c5efa0190469791dd3a105bebd81b2891596683001c6cb0d7446fc0d5125a2", "enabled": 1 } }, "privoxy": { "100": { "checksum": "sha256:0c0a1216baa5adcd33f85d92f248893c11cb55b5b068db990b2f94d9d7914932", "enabled": 1 } }, "procmail": { "100": { "checksum": "sha256:ab3962f174f9a9a125e2aad039584a83294b11d3ee56c8b477217a8193afc87d", "enabled": 1 } }, "prosody": { "100": { "checksum": "sha256:3573621c7a830f0d511041287fe8bb073eaa5e200c5b8b2e72096bf469c35f1d", "enabled": 1 } }, "psad": { "100": { "checksum": "sha256:72c77f99e96935aa093669bb095f914484b67e54ae66b4ef8ca780d2758cc6c8", "enabled": 1 } }, "ptchown": { "100": { "checksum": "sha256:318c62c2f63baefa43a61f8ef58bc168596087cd892a63afee72b1cb29cac0c7", "enabled": 1 } }, "pulseaudio": { "100": { "checksum": "sha256:3a7087d4680d1a35893caebf703937f36c1b1799dabc446b02428afc6c163c6f", "enabled": 1 } }, "puppet": { "100": { "checksum": "sha256:366806c68d03db1138cc9a3a08e0fdc4883825c20d1bed9a8e64ce7662bfb032", "enabled": 1 } }, "pwauth": { "100": { "checksum": "sha256:d23c293e39c1a025d375b714dd9319d27545c6c62b58342ef530763b84a24b16", "enabled": 1 } }, "qatlib": { "100": { "checksum": "sha256:cef00aa0bebb89745286c7db05374eb06fe9637c480d295f80f31d21be53672f", "enabled": 1 } }, "qmail": { "100": { "checksum": "sha256:5617289b46ae349fafd315cf92082d64119c6a0b5606bf79aa17fc99e34985b6", "enabled": 1 } }, "qpid": { "100": { "checksum": "sha256:1ee83ce320b057b0569c4bc9f8abdab6233eaf6fcd9c1a78975040bb2d9ddc56", "enabled": 1 } }, "quantum": { "100": { "checksum": "sha256:704e2c7d5fca4cb84e320b9359c6168037ebc87908330c0053c143c8821d3863", "enabled": 1 } }, "quota": { "100": { "checksum": "sha256:58dca8ea36c665ab74eb79cd51d54a7224514660bdfc4b666c2b8031e2877da3", "enabled": 1 } }, "rabbitmq": { "100": { "checksum": "sha256:acc51303259d5e6049265839c5ed13703baa80ae3477ae292103b4665c83dfbb", "enabled": 1 } }, "radius": { "100": { "checksum": "sha256:b53cd6e9f7493004e2b5bdf30e368fbc90b4f971db3fe56e55f31cc839969237", "enabled": 1 } }, "radvd": { "100": { "checksum": "sha256:c909b5d3b8ba285d71358dc7c9fca270e7b1586266c597a60b167ff600d10f7b", "enabled": 1 } }, "raid": { "100": { "checksum": "sha256:071d5ed18ad1c7d334bfba17d392c80a588ad00b1052ab0e21f9c53b0db95e4f", "enabled": 1 } }, "rasdaemon": { "100": { "checksum": "sha256:6d5c3ce954d1cef655ba323f1fef77bda8813815c0d991d020e0c760a4089aa5", "enabled": 1 } }, "rdisc": { "100": { "checksum": "sha256:3b10e49a0fec5da0e8998d77835464b25ced6ee9bd04a24dd861601a9aa02d2d", "enabled": 1 } }, "readahead": { "100": { "checksum": "sha256:e5b3d0b401606c60c1acdfce14c7b6b3f3363a1f58d8d91dd3be36e3df38005f", "enabled": 1 } }, "realmd": { "100": { "checksum": "sha256:b37f4de6bcd25f24eee675c18e2f1842023f0cf9a9a9a7923b64f7eaa875494f", "enabled": 1 } }, "redis": { "100": { "checksum": "sha256:b862f723b5285ad797d5a7546a17b1080690d9218a23b84bb0f7fe618169610f", "enabled": 1 } }, "remotelogin": { "100": { "checksum": "sha256:622d5ae5349b504c7ec9e472a86c5a1e1bc27606868d53d154112a63299c8dc9", "enabled": 1 } }, "restraint": { "400": { "checksum": "sha256:95d0d03fbc1d4147f02a0b3da7cd76efbdd75d1f5812cf6160e922336abbf270", "enabled": 1 } }, "rhcd": { "100": { "checksum": "sha256:fcd8c8415a58d937e726318252127c28e69af2819be01b977b4dda03994834a0", "enabled": 1 } }, "rhcs": { "100": { "checksum": "sha256:1e8298e1e453be52e1a86e391b827f2c71729c066f7d6c63abe7843e28647207", "enabled": 1 } }, "rhev": { "100": { "checksum": "sha256:eaa7e5657902e356ab7700656ff4c185eaf8df675d0a9a3f9ea1a4d8d8fb577d", "enabled": 1 } }, "rhgb": { "100": { "checksum": "sha256:5271469985f35fb7911a72e9e705bc35971294386838f1f11033657874821cea", "enabled": 1 } }, "rhnsd": { "100": { "checksum": "sha256:2a502ce16b7c56df48d0798ed7ca60370d0993ee425a542bf0621b51d4caf2d4", "enabled": 1 } }, "rhsmcertd": { "100": { "checksum": "sha256:80b01c4c3ddf665dd2fd5391b6069e9b868573f7c7a9d4421bc5bfeaadc71a22", "enabled": 1 } }, "rhts": { "400": { "checksum": "sha256:39f15722a115b11064c20b8bc2758e8fe06531a8f923ea00a3e7079a0554e3d6", "enabled": 1 } }, "ricci": { "100": { "checksum": "sha256:dd355c8a0a048baee894e0951365b525bbf5d7cce5907c9919bb0fffb1876661", "enabled": 1 } }, "rkhunter": { "100": { "checksum": "sha256:95ecb10dfe02e0384b0f220c46a7d0c85166a4e2cca0bb55654420b73611a646", "enabled": 1 } }, "rlogin": { "100": { "checksum": "sha256:d8594a73f246468d42569485bb148cbf1b582779074735f6da46e92f763984e1", "enabled": 1 } }, "rngd": { "100": { "checksum": "sha256:50b57a4776d2c6de8b1698bd31ecbfcc98d90d4a9bd2c95672b114c9637ce940", "enabled": 1 } }, "roundup": { "100": { "checksum": "sha256:54c5e85363d326f78c5896bf87c36fca9bc28acee3bd9da4d6d22f090dc9f7c6", "enabled": 1 } }, "rpc": { "100": { "checksum": "sha256:404f69ae03180ddce44c15efc636e5d7ff34b2108cb2c09624075be42a0545b1", "enabled": 1 } }, "rpcbind": { "100": { "checksum": "sha256:f5c703b4107b2616b1adf105b5c592c79f03df1f3a24f31453a242cdec50ea97", "enabled": 1 } }, "rpm": { "100": { "checksum": "sha256:bb52185531861597a0b44bca75497c08ac4c07e10ec63505c333bde800176fb7", "enabled": 1 } }, "rrdcached": { "100": { "checksum": "sha256:2cb11286a0afa94eccbd33de5a32b16d9e178e41ac1fec9dcd801b77dd257051", "enabled": 1 } }, "rshd": { "100": { "checksum": "sha256:0052600f999e54f224d1210c5920b42227cd98ce8db24155b0cfecdf4f6c6474", "enabled": 1 } }, "rshim": { "100": { "checksum": "sha256:7825528bbcdb081ead1d0d2da0d5156e430f93d003f58dcebbb8997c906f0dec", "enabled": 1 } }, "rssh": { "100": { "checksum": "sha256:cde9bfc64b345835e9f8012426cf218bad9972c70e22e0ef6d6a90cc5fa7163c", "enabled": 1 } }, "rsync": { "100": { "checksum": "sha256:1295d1a5434aa1df7fd0f0e2a9b2b045b0ec51bd8619d528349530b322eddf24", "enabled": 1 } }, "rtas": { "100": { "checksum": "sha256:8c5c3a495b42ed14f56e360dd9eaaa0df2426f2c93676880f129bab95e247338", "enabled": 1 } }, "rtkit": { "100": { "checksum": "sha256:7b0c8b0fd23723aed98bc581f2ead99da87893c6ba042ec19ca731e1d802eab6", "enabled": 1 } }, "rwho": { "100": { "checksum": "sha256:fe02b4fb75b8a75ac77b53ae13939199422543f68e71829fd1530aa34c8715cc", "enabled": 1 } }, "samba": { "100": { "checksum": "sha256:b0af5e99697f8e864538813653ea879ea50b6a9269424ca1deaa95ec1a4a7778", "enabled": 1 } }, "sambagui": { "100": { "checksum": "sha256:cde09885f6d657107c517f09f01933816199f0694799fe3df4ac281b411edbb9", "enabled": 1 } }, "sandboxX": { "100": { "checksum": "sha256:ffd79700483dc635b0396df2f9ca75062aff985e7874c2014bca95e80b08bd5d", "enabled": 1 } }, "sanlock": { "100": { "checksum": "sha256:fcd9cea4d61e3597a5e5cb1bd36efc228a21e5b8ee79c68b44e8a93aae834fb6", "enabled": 1 } }, "sap": { "100": { "checksum": "sha256:4cffd87a27c2b48cc59f596a42e860ea52acdfcbff0c2a1e4aeafa3338b29e6e", "enabled": 1 } }, "sasl": { "100": { "checksum": "sha256:ed40cd297b033fd86770692e7d3ab76598684f521f079475d8b44953116291cf", "enabled": 1 } }, "sbd": { "100": { "checksum": "sha256:8c40daa2d8288e7535e9db7724d16a34add080c65fab32bdf5e10cdace95a063", "enabled": 1 } }, "sblim": { "100": { "checksum": "sha256:73b934662023f728bfeb5cac500547e15e389bad3575a3750bf05901d4396531", "enabled": 1 } }, "screen": { "100": { "checksum": "sha256:8dcba8c325fdb2c56f7c7c77362e287469484b19f0b491dd7bc0abc6e033378b", "enabled": 1 } }, "secadm": { "100": { "checksum": "sha256:2d9e07e5ee506dbf6cf976e257210fdb14222c1e437b1f7dceb1548fa399beba", "enabled": 1 } }, "sectoolm": { "100": { "checksum": "sha256:e6c4cabc853877f459381f06d4d47660ce9360a188ab87bb825f1f1eaed39d7d", "enabled": 1 } }, "selinuxutil": { "100": { "checksum": "sha256:38da361f759758e151b7f8612954c91c0fde09615e1c82242c80706231e16355", "enabled": 1 } }, "sendmail": { "100": { "checksum": "sha256:286d18507d80fcfa7cb83ed21321df23c3a29001777890137c1b52c113cd8489", "enabled": 1 } }, "sensord": { "100": { "checksum": "sha256:6681fdfffe9ba81fa3712f665b98feb68f9331f8f282faf19a4ae5e70b02bd63", "enabled": 1 } }, "setrans": { "100": { "checksum": "sha256:69a3bce8a9e35afa4c3b4c7e2f4ba604491313a688d36e3158f96e90343ea370", "enabled": 1 } }, "setroubleshoot": { "100": { "checksum": "sha256:c0d9c0c6cea100697cbc684f2ab74ec66d7c121ca88200752585aeaf4c2b2596", "enabled": 1 } }, "seunshare": { "100": { "checksum": "sha256:219912d16220957b98bbe694f06c3509d2f884ff8418da1eeeda43f4dce81dc9", "enabled": 1 } }, "shorewall": { "100": { "checksum": "sha256:d0dd25c1a5b77a28fcf61f4796c479a026a8e2ec944f5e4d1492cd552aa9ca02", "enabled": 1 } }, "slocate": { "100": { "checksum": "sha256:8d149bca3982ac8c0f7376571d1f0614f05eb2fe7465951eefb17153f642ee40", "enabled": 1 } }, "slpd": { "100": { "checksum": "sha256:6798797c2dda6b8b24aada5a4bbc4f8775faca22435ed1de39193457043f8d8b", "enabled": 1 } }, "smartmon": { "100": { "checksum": "sha256:9de88cb8f60049e56dc5ead706a324189ff276929ccc3bf4ba7d06e96c0d8414", "enabled": 1 } }, "smokeping": { "100": { "checksum": "sha256:29560964ba6e3cb5fc392417ff4aa3e50a5f15c2bc21bdf37088c87868ffeec7", "enabled": 1 } }, "smoltclient": { "100": { "checksum": "sha256:0e47cf6e775a1e2ecb506e444731e086d057b7f2c2aeb6bce7b636372de88d95", "enabled": 1 } }, "snapper": { "100": { "checksum": "sha256:e264d81b3e1173c98c1efc90529d7f9c2c9047b843e29a4fb807d0b1b96b4685", "enabled": 1 } }, "snmp": { "100": { "checksum": "sha256:ec6638b69ec04510a807a5d2c97568a670f80a7999ee711e5e1874136923718e", "enabled": 1 } }, "snort": { "100": { "checksum": "sha256:67c1751c124c249fa15c36f948b60a76db3729f819678cecba569c79abec3af2", "enabled": 1 } }, "sosreport": { "100": { "checksum": "sha256:345531e273b1a0c0d6e02a1898d60759007d5126ec2016eea320832c33ac8211", "enabled": 1 } }, "soundserver": { "100": { "checksum": "sha256:932d162b8365f213b97eb6a7e62d20b8c30bfe1863b6534462da96077d6347fc", "enabled": 1 } }, "spamassassin": { "100": { "checksum": "sha256:fbceb12b459776c22867efa6bf2d07657e058c4a80d6f20e724086ec58e600b0", "enabled": 1 } }, "speech-dispatcher": { "100": { "checksum": "sha256:092593b379ff3e7b3d15b0dfdd2adfde73fb76401d1b6b6e79a4efccbf03f0c9", "enabled": 1 } }, "squid": { "100": { "checksum": "sha256:4374a853cb79bd24d05f65585ad1a74a7b9da4d3454ca233da351e009e75ed79", "enabled": 1 } }, "ssh": { "100": { "checksum": "sha256:493ee98f9cda10ca995eb5600d2940556d70cb3f544501c6bd1968745cc234fa", "enabled": 1 } }, "sslh": { "100": { "checksum": "sha256:5dc03f6d33a6caa7c2fb5ede0f8ded4ecb8eb244443614189cb5909521cab539", "enabled": 1 } }, "sssd": { "100": { "checksum": "sha256:4ee157395107db3b146239f14c9868f6e9d71199520e5d2e473e5e836d0635d1", "enabled": 1 } }, "staff": { "100": { "checksum": "sha256:6a89d52653ce210b1dfb2d2b10e184cacb18976d238fc92b349c40661fdcbf47", "enabled": 1 } }, "stalld": { "100": { "checksum": "sha256:14aa28921885421f92fb51920ca0d8fde122881ac9e28d024c80a7b234b82dd3", "enabled": 1 } }, "stapserver": { "100": { "checksum": "sha256:1610ffc61426c0e791bd7a6a2628935edf3d66fc7818fb42c314044a26ac5d2d", "enabled": 1 } }, "stratisd": { "100": { "checksum": "sha256:5025b28ad7090d4802078c92ba257b46909bbb74fcd6cae0f2a5d65978a5ff8e", "enabled": 1 } }, "stunnel": { "100": { "checksum": "sha256:67b1c960e0cd11cdf1446615cf633700585e69f41c4039cfb183d69527648f1e", "enabled": 1 } }, "su": { "100": { "checksum": "sha256:60850779bd04d036ea792653d4e794f0e865936648c03030515d8125cf412b4b", "enabled": 1 } }, "sudo": { "100": { "checksum": "sha256:0965d3f30b2813a4364e9ce8da0f8916579319ed9507640383c3bf195ae1c00e", "enabled": 1 } }, "svnserve": { "100": { "checksum": "sha256:d589d2e8bddbcea03272bb589cffde776ccb410ac685e5071896fa8e59660c7f", "enabled": 1 } }, "swift": { "100": { "checksum": "sha256:407ddd533381c1f822ec8471e639404c4b926e7e635b3d25d4ba1403a5c517b3", "enabled": 1 } }, "sysadm": { "100": { "checksum": "sha256:1d310d8df59d431579fa3568b7ac14f853223aadd4087fe4c57a1add6606987d", "enabled": 1 } }, "sysadm_secadm": { "100": { "checksum": "sha256:6a0a3b3c4028f304beeacb7d06c62975a2bfb13d67a6a4fe6b95f6d6eeaa74b7", "enabled": 1 } }, "sysnetwork": { "100": { "checksum": "sha256:56b999cfaf7defa2c5cdd5c0c16e7c9c1de940dcc528778ff752fca51f95b38a", "enabled": 1 } }, "sysstat": { "100": { "checksum": "sha256:358b9c77d350c4bc17cf773e99f81d5267f8c41d47402234be37dd9f31655ad0", "enabled": 1 } }, "systemd": { "100": { "checksum": "sha256:124460467ec0f3f7443a67583ef4613154b2b92615e0d779a3ec8060c5a8ce4a", "enabled": 1 } }, "tangd": { "100": { "checksum": "sha256:a2e2a5d807eabe30a09085dcec38d530781c1f02e840a6ed02ccc32bd2738e60", "enabled": 1 } }, "targetd": { "100": { "checksum": "sha256:4a9e67698b4759a406188562c375ecd847f205c38873fb895c3c20abc8bac869", "enabled": 1 } }, "tcpd": { "100": { "checksum": "sha256:b69263d84bf511825be889b667401c30964c7d30d1147d577ae294cd7f05aa01", "enabled": 1 } }, "tcsd": { "100": { "checksum": "sha256:fa40e623b09fb5f61c2c2113a8ed0d17a4a5d105b9f22389ce088a766c0042c0", "enabled": 1 } }, "telepathy": { "100": { "checksum": "sha256:d98021b3bbdd6255c2009e5adafbf41faf3370434396cf2e928e7df693dd50b6", "enabled": 1 } }, "telnet": { "100": { "checksum": "sha256:abf90c09c67553f2b04cab7461138bfd8689752a752a79e36e248f209eae73e6", "enabled": 1 } }, "tftp": { "100": { "checksum": "sha256:f5dcde99c32caa66ea23a9be3b2d6d95550eff7db62b11733f16f1f5eb01bce2", "enabled": 1 } }, "tgtd": { "100": { "checksum": "sha256:6a96f6d7abaa0e1951832347e924c5aff5c66194da3881ec4855e1401a5ce6d2", "enabled": 1 } }, "thin": { "100": { "checksum": "sha256:b0c1055c6a8419035ee9a9a45de2bcdd7183a906bb40d98bb2be80b882a60c51", "enabled": 1 } }, "thumb": { "100": { "checksum": "sha256:928596446716a5d1ef1f8958d150959e59a6258975ebf453c2d5a2747e6d59fd", "enabled": 1 } }, "tlp": { "100": { "checksum": "sha256:540f0a0339feec1e11d145c6ba99f700750451a2fad43190e256e590c3835950", "enabled": 1 } }, "tmpreaper": { "100": { "checksum": "sha256:4370bc58da3c779088151878e7291e03fcd2ed706425b4f2eeeed7097f8adcad", "enabled": 1 } }, "tomcat": { "100": { "checksum": "sha256:98e40c398d94d82eea102055d52f59f77581955b0e463dd95386b4559a4cc852", "enabled": 1 } }, "tor": { "100": { "checksum": "sha256:b309ee2c391d91ace45662d2152870961635ca8968487d0f2570508e160dbd9b", "enabled": 1 } }, "tuned": { "100": { "checksum": "sha256:90bcbb5cad47c032a02390e0c4b22ff79f93fbfd2eb4a385441a87532d6aa632", "enabled": 1 } }, "tvtime": { "100": { "checksum": "sha256:a4438b3063cc1760c663e35a7be70a7dcd099de505ce1562e02090ac6b635787", "enabled": 1 } }, "udev": { "100": { "checksum": "sha256:f308ba2a718171dc871e2586f35a90cb1bbc37decf13e714af72b38cc44505a2", "enabled": 1 } }, "ulogd": { "100": { "checksum": "sha256:05aa3f423e0505dfef9b5e311b0f0a71d93159c7f6440979d3c00a5d9d22ef4c", "enabled": 1 } }, "uml": { "100": { "checksum": "sha256:e0257dd4e250c5aa0baa8151d48942735949bd3018318c11c28a4a5df2aacb90", "enabled": 1 } }, "unconfined": { "100": { "checksum": "sha256:26c4304a314b04f6269b45b4285b449caad98d22de7b3017ef648212c6674da1", "enabled": 1 } }, "unconfineduser": { "100": { "checksum": "sha256:af2778eed6f3cc25c5390e4b1a892995c37a83f5952f51820007b8cd387833b4", "enabled": 1 } }, "unlabelednet": { "100": { "checksum": "sha256:c1de2adfab2cd577a0cc9f2c1a50cd6321324d31b9c3f81be9231f85d4acbbd3", "enabled": 1 } }, "unprivuser": { "100": { "checksum": "sha256:dc1ddbf10a94d55c476ff343c862c24dc649d82190a16a0fd924c527743a667d", "enabled": 1 } }, "updfstab": { "100": { "checksum": "sha256:11eb931e80afefafd0f1c41ea19add0c746170447c76dd629146c43bd9e98067", "enabled": 1 } }, "usbmodules": { "100": { "checksum": "sha256:3a8334855e0cfe67e2c29198baacf61ea0e397d5da30f60fb844044cd105557e", "enabled": 1 } }, "usbmuxd": { "100": { "checksum": "sha256:e883ad3abf2e2792abf4122f99db904139b2f1833b7d99c45b34a7c8244d7c66", "enabled": 1 } }, "userdomain": { "100": { "checksum": "sha256:b6cc74d18c0d17741092aaa8e830f160021ad38bdc38880e3b1488ef30f59a49", "enabled": 1 } }, "userhelper": { "100": { "checksum": "sha256:0a80be6b1635de7593f46421219be5a6ccb9c22973c60b74436ba16884d789fc", "enabled": 1 } }, "usermanage": { "100": { "checksum": "sha256:c56dfa32f163d476dd5761e762ea743189a97636acaf2b4eefcf88812e624877", "enabled": 1 } }, "usernetctl": { "100": { "checksum": "sha256:9c0430aca661392960008075f0a1a934f0c05239976e73c4f0408e9cb5e279ca", "enabled": 1 } }, "uucp": { "100": { "checksum": "sha256:527db920542a250b24e24cb8c6bd86502c3be7e270d11f43d4ecb1a6632f8421", "enabled": 1 } }, "uuidd": { "100": { "checksum": "sha256:b3df65ab2d82cbb05e872f00f46156cfbae937a3a6f850d05f43f4866b772af8", "enabled": 1 } }, "varnishd": { "100": { "checksum": "sha256:9174c3c68bf6b4c2b2b9be7e2a0d985ce6edfdec0dc4c225b427b986449d3e7a", "enabled": 1 } }, "vdagent": { "100": { "checksum": "sha256:52636260af5ae220698d505c6a90a33a71f5878a2cea3a303033c896f000b732", "enabled": 1 } }, "vhostmd": { "100": { "checksum": "sha256:df2d64343b871c840603e82f5fc1f7928c0345b314450e536f9ef66becf131ad", "enabled": 1 } }, "virt": { "100": { "checksum": "sha256:c209f246ed99c66defe424c0f1bd13e970df31d84d2f4e4a2a79ef1c00c3bbe2", "enabled": 1 } }, "virt_supplementary": { "100": { "checksum": "sha256:e8147b1556404b4a2ce521cc4c22e28d4ee42045ea2480043fc8bd2d5a593207", "enabled": 1 } }, "vlock": { "100": { "checksum": "sha256:54a1b28cebc1928622edc4f1f1762a5f2d57783be153052d592f2d10dd9b24a1", "enabled": 1 } }, "vmtools": { "100": { "checksum": "sha256:323b80c9ba1efb9852725d635a5aa442ae2bbd4ec74360cf75853616401dd4f9", "enabled": 1 } }, "vmware": { "100": { "checksum": "sha256:b68b94c1c4662f7005faac011c887619dbb76d69d3d2b63f354d48a144d0d63c", "enabled": 1 } }, "vnstatd": { "100": { "checksum": "sha256:c544da94ee903d411f7ab352b3fe286fa0a1b0fb8490d508565df66d5501b852", "enabled": 1 } }, "vpn": { "100": { "checksum": "sha256:035f214c319dcee2ce4e9751fa9cf1343d81d569409f3a56faf79e506e445504", "enabled": 1 } }, "w3c": { "100": { "checksum": "sha256:d6c6c226dcc92d9a668f42b86dde398d2d875b479915af656734357e20ed0e96", "enabled": 1 } }, "watchdog": { "100": { "checksum": "sha256:a237aeaf94bed19de41ef5289be96ce5bf85dce6f2fc4f9d36d6185b4e59520a", "enabled": 1 } }, "wdmd": { "100": { "checksum": "sha256:1012fdefa5dcf2e540f536de5da79df0ba715ccc2a86a414e2e846675d32495f", "enabled": 1 } }, "webadm": { "100": { "checksum": "sha256:f374fcd23ed5f30ce1159470af1e1254e0352a6ad7cf261287de8ec51f2b2626", "enabled": 1 } }, "webalizer": { "100": { "checksum": "sha256:b66dfdfd23ed304d5f416bb0d8de4565d28cd732253fc2f9e4b4bac59515790f", "enabled": 1 } }, "wine": { "100": { "checksum": "sha256:a4a3be2a0feea4de24ca26c0eed58c7025304d810ee9d4a3e1b7e964a8d28c6d", "enabled": 1 } }, "wireguard": { "100": { "checksum": "sha256:3084b8fd20df45b718cb6779240f4292a0437f536bde7db3616c40b1500ba0f6", "enabled": 1 } }, "wireshark": { "100": { "checksum": "sha256:3ec8e55883a8b73de513f533ed80b1b665e3ed650fdf2b471e5565e1cc3034d9", "enabled": 1 } }, "xen": { "100": { "checksum": "sha256:adabf99dccce9182074a23fb8f05ac537899170fbbfe022f39cb520275339186", "enabled": 1 } }, "xguest": { "100": { "checksum": "sha256:8a7f7d9514fa5257088e451dcf96c6c221a7b6499cae9ac15a688cbe2e84def9", "enabled": 1 } }, "xserver": { "100": { "checksum": "sha256:ca8b72b391320d267580cebf2ad9ef811a781b34647e1c63e525e982911cf956", "enabled": 1 } }, "zabbix": { "100": { "checksum": "sha256:ed9800f41440cf137baa7864acdd8a503b7e051cf023d252b87daeb5357be889", "enabled": 1 } }, "zarafa": { "100": { "checksum": "sha256:c3a9d1c95f377fafb9e94a84bec89645a1053d23b557b143367992eae9993596", "enabled": 1 } }, "zebra": { "100": { "checksum": "sha256:77d57a56a3196f09ea5eb808f8c3b4cef333b241e9007b91779ba78a857e5cfa", "enabled": 1 } }, "zoneminder": { "100": { "checksum": "sha256:675bdf525f863eaf17fe6c25507cf2dc3c4ab05d1e4b0d3477573d34980a5cd2", "enabled": 1 } }, "zosremote": { "100": { "checksum": "sha256:79bb5490f9f16e8caff67fc87a834bf6f11647cfed990c421616eb1adaac5c70", "enabled": 1 } } }, "selinux_priorities": true }, "changed": false } TASK [fedora.linux_system_roles.selinux : Load SELinux modules] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:115 Saturday 11 January 2025 11:30:33 -0500 (0:00:03.226) 0:01:58.734 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_modules is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Restore SELinux labels on filesystem tree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:128 Saturday 11 January 2025 11:30:33 -0500 (0:00:00.051) 0:01:58.786 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Restore SELinux labels on filesystem tree in check mode] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:136 Saturday 11 January 2025 11:30:33 -0500 (0:00:00.040) 0:01:58.826 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:153 Saturday 11 January 2025 11:30:33 -0500 (0:00:00.061) 0:01:58.888 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:157 Saturday 11 January 2025 11:30:33 -0500 (0:00:00.039) 0:01:58.927 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:166 Saturday 11 January 2025 11:30:33 -0500 (0:00:00.034) 0:01:58.962 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:175 Saturday 11 January 2025 11:30:34 -0500 (0:00:00.039) 0:01:59.001 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:182 Saturday 11 January 2025 11:30:34 -0500 (0:00:00.044) 0:01:59.046 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml for managed-node2 => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:14 Saturday 11 January 2025 11:30:34 -0500 (0:00:00.182) 0:01:59.228 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_spec": { "state": "absent" }, "__podman_kube_str": "apiVersion: v1\nkind: Pod\nmetadata:\n labels:\n app: test\n io.containers.autoupdate: registry\n name: bogus\nspec:\n containers:\n - name: bogus\n image: quay.io/libpod/testimage:20210610\n" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:21 Saturday 11 January 2025 11:30:34 -0500 (0:00:00.265) 0:01:59.493 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_kube": { "apiVersion": "v1", "kind": "Pod", "metadata": { "labels": { "app": "test", "io.containers.autoupdate": "registry" }, "name": "bogus" }, "spec": { "containers": [ { "image": "quay.io/libpod/testimage:20210610", "name": "bogus" } ] } }, "__podman_kube_file": "", "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:33 Saturday 11 January 2025 11:30:34 -0500 (0:00:00.092) 0:01:59.585 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_name": "bogus", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:38 Saturday 11 January 2025 11:30:34 -0500 (0:00:00.070) 0:01:59.655 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 11 January 2025 11:30:34 -0500 (0:00:00.126) 0:01:59.782 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 11 January 2025 11:30:34 -0500 (0:00:00.072) 0:01:59.854 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 11 January 2025 11:30:34 -0500 (0:00:00.080) 0:01:59.935 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Saturday 11 January 2025 11:30:35 -0500 (0:00:00.097) 0:02:00.032 ****** ok: [managed-node2] => { "changed": false, "stat": { "atime": 1736612945.6352425, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1736612940.9742577, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 9160785, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "1643853349", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Saturday 11 January 2025 11:30:35 -0500 (0:00:00.443) 0:02:00.475 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Saturday 11 January 2025 11:30:35 -0500 (0:00:00.086) 0:02:00.562 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Saturday 11 January 2025 11:30:35 -0500 (0:00:00.068) 0:02:00.630 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Saturday 11 January 2025 11:30:35 -0500 (0:00:00.067) 0:02:00.698 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Saturday 11 January 2025 11:30:35 -0500 (0:00:00.143) 0:02:00.842 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Saturday 11 January 2025 11:30:35 -0500 (0:00:00.071) 0:02:00.913 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Saturday 11 January 2025 11:30:35 -0500 (0:00:00.071) 0:02:00.985 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Saturday 11 January 2025 11:30:36 -0500 (0:00:00.072) 0:02:01.057 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if no kube spec is given] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:43 Saturday 11 January 2025 11:30:36 -0500 (0:00:00.069) 0:02:01.127 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_kube", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:52 Saturday 11 January 2025 11:30:36 -0500 (0:00:00.065) 0:02:01.192 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:60 Saturday 11 January 2025 11:30:36 -0500 (0:00:00.099) 0:02:01.291 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_path": "/etc/containers/ansible-kubernetes.d" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:64 Saturday 11 January 2025 11:30:36 -0500 (0:00:00.050) 0:02:01.342 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_file": "/etc/containers/ansible-kubernetes.d/bogus.yml" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:68 Saturday 11 January 2025 11:30:36 -0500 (0:00:00.050) 0:02:01.393 ****** ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Get service name using systemd-escape] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:75 Saturday 11 January 2025 11:30:36 -0500 (0:00:00.054) 0:02:01.448 ****** ok: [managed-node2] => { "changed": false, "cmd": [ "systemd-escape", "--template", "podman-kube@.service", "/etc/containers/ansible-kubernetes.d/bogus.yml" ], "delta": "0:00:00.005692", "end": "2025-01-11 11:30:36.777443", "rc": 0, "start": "2025-01-11 11:30:36.771751" } STDOUT: podman-kube@-etc-containers-ansible\x2dkubernetes.d-bogus.yml.service TASK [fedora.linux_system_roles.podman : Cleanup containers and services] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:83 Saturday 11 January 2025 11:30:36 -0500 (0:00:00.416) 0:02:01.864 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:2 Saturday 11 January 2025 11:30:37 -0500 (0:00:00.208) 0:02:02.073 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:10 Saturday 11 January 2025 11:30:37 -0500 (0:00:00.065) 0:02:02.138 ****** ok: [managed-node2] => { "changed": false, "enabled": false, "failed_when_result": false, "name": "podman-kube@-etc-containers-ansible\\x2dkubernetes.d-bogus.yml.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "network-online.target systemd-journald.socket \"system-podman\\\\x2dkube.slice\" sysinit.target basic.target -.mount", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BindLogSockets": "no", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanLiveMount": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DebugInvocation": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "A template for running K8s workloads via podman-kube-play", "DevicePolicy": "auto", "Documentation": "\"man:podman-kube-play(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698229248", "EffectiveMemoryMax": "3698229248", "EffectiveTasksMax": "22365", "Environment": "\"PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\\\x2dkubernetes.d-bogus.yml.service\"", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true /etc/containers/ansible-kubernetes.d/bogus.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true /etc/containers/ansible-kubernetes.d/bogus.yml ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/ansible-kubernetes.d/bogus.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/ansible-kubernetes.d/bogus.yml ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/podman-kube@.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "podman-kube@-etc-containers-ansible\\x2dkubernetes.d-bogus.yml.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13978", "LimitNPROCSoft": "13978", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13978", "LimitSIGPENDINGSoft": "13978", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LiveMountResult": "success", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureDurationUSec": "[not set]", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3175989248", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "[not set]", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "[not set]", "MemorySwapMax": "infinity", "MemorySwapPeak": "[not set]", "MemoryZSwapCurrent": "[not set]", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"podman-kube@-etc-containers-ansible\\\\x2dkubernetes.d-bogus.yml.service\"", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivatePIDs": "no", "PrivateTmp": "no", "PrivateTmpEx": "no", "PrivateUsers": "no", "PrivateUsersEx": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectControlGroupsEx": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "-.mount sysinit.target \"system-podman\\\\x2dkube.slice\"", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system-podman\\x2dkube.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22365", "TimeoutAbortUSec": "1min 10s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 10s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-online.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.podman : Check if kube file exists] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:28 Saturday 11 January 2025 11:30:37 -0500 (0:00:00.615) 0:02:02.753 ****** ok: [managed-node2] => { "changed": false, "stat": { "atime": 1736612991.5220592, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "f8266a972ed3be7e204d2a67883fe3a22b8dbf18", "ctime": 1736612990.9540617, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 343933123, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1736612990.671063, "nlink": 1, "path": "/etc/containers/ansible-kubernetes.d/bogus.yml", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 237, "uid": 0, "version": "822373163", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Remove pod/containers] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:38 Saturday 11 January 2025 11:30:38 -0500 (0:00:00.439) 0:02:03.193 ****** changed: [managed-node2] => { "actions": [ "/usr/bin/podman kube play --down /etc/containers/ansible-kubernetes.d/bogus.yml" ], "changed": true, "failed_when_result": false } STDOUT: Pods stopped: 6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7 Pods removed: 6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7 Secrets removed: Volumes removed: TASK [fedora.linux_system_roles.podman : Remove kubernetes yaml file] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:56 Saturday 11 January 2025 11:30:38 -0500 (0:00:00.605) 0:02:03.799 ****** changed: [managed-node2] => { "changed": true, "path": "/etc/containers/ansible-kubernetes.d/bogus.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:61 Saturday 11 January 2025 11:30:39 -0500 (0:00:00.416) 0:02:04.215 ****** changed: [managed-node2] => { "changed": true, "cmd": [ "podman", "image", "prune", "-f" ], "delta": "0:00:00.032448", "end": "2025-01-11 11:30:39.602635", "rc": 0, "start": "2025-01-11 11:30:39.570187" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_kube_spec.yml:70 Saturday 11 January 2025 11:30:39 -0500 (0:00:00.477) 0:02:04.693 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 11 January 2025 11:30:39 -0500 (0:00:00.169) 0:02:04.862 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 11 January 2025 11:30:39 -0500 (0:00:00.046) 0:02:04.908 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 11 January 2025 11:30:39 -0500 (0:00:00.043) 0:02:04.951 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update containers and services] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:87 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.117) 0:02:05.069 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:189 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.039) 0:02:05.108 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:196 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.037) 0:02:05.145 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle credential files - absent] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:202 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.034) 0:02:05.180 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:211 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.035) 0:02:05.216 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [Run role] **************************************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:291 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.057) 0:02:05.274 ****** included: fedora.linux_system_roles.podman for managed-node2 TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.093) 0:02:05.368 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.062) 0:02:05.431 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.045) 0:02:05.476 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.102) 0:02:05.579 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.057) 0:02:05.637 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.060) 0:02:05.697 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.072) 0:02:05.770 ****** ok: [managed-node2] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed-node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node2] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } ok: [managed-node2] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Saturday 11 January 2025 11:30:40 -0500 (0:00:00.170) 0:02:05.941 ****** ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Saturday 11 January 2025 11:30:41 -0500 (0:00:00.924) 0:02:06.865 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Saturday 11 January 2025 11:30:41 -0500 (0:00:00.069) 0:02:06.935 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages))", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Saturday 11 January 2025 11:30:42 -0500 (0:00:00.085) 0:02:07.020 ****** skipping: [managed-node2] => { "false_condition": "__podman_is_transactional | d(false)" } TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Saturday 11 January 2025 11:30:42 -0500 (0:00:00.046) 0:02:07.067 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Saturday 11 January 2025 11:30:42 -0500 (0:00:00.046) 0:02:07.113 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Saturday 11 January 2025 11:30:42 -0500 (0:00:00.096) 0:02:07.210 ****** ok: [managed-node2] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.028902", "end": "2025-01-11 11:30:42.560414", "rc": 0, "start": "2025-01-11 11:30:42.531512" } STDOUT: podman version 5.3.1 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Saturday 11 January 2025 11:30:42 -0500 (0:00:00.444) 0:02:07.654 ****** ok: [managed-node2] => { "ansible_facts": { "podman_version": "5.3.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Saturday 11 January 2025 11:30:42 -0500 (0:00:00.075) 0:02:07.730 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Saturday 11 January 2025 11:30:42 -0500 (0:00:00.074) 0:02:07.804 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "(podman_quadlet_specs | length > 0) or (podman_secrets | length > 0)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Saturday 11 January 2025 11:30:42 -0500 (0:00:00.073) 0:02:07.878 ****** META: end_host conditional evaluated to False, continuing execution for managed-node2 skipping: [managed-node2] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node2" } MSG: end_host conditional evaluated to false, continuing execution for managed-node2 TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Saturday 11 January 2025 11:30:42 -0500 (0:00:00.089) 0:02:07.967 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__has_type_pod or __has_pod_file_ext or __has_pod_file_src_ext or __has_pod_template_src_ext or __has_pod_template_src_ext_j2", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:96 Saturday 11 January 2025 11:30:43 -0500 (0:00:00.089) 0:02:08.057 ****** META: end_host conditional evaluated to False, continuing execution for managed-node2 skipping: [managed-node2] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node2" } MSG: end_host conditional evaluated to false, continuing execution for managed-node2 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:109 Saturday 11 January 2025 11:30:43 -0500 (0:00:00.123) 0:02:08.180 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 11 January 2025 11:30:43 -0500 (0:00:00.130) 0:02:08.311 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 11 January 2025 11:30:43 -0500 (0:00:00.151) 0:02:08.463 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 11 January 2025 11:30:43 -0500 (0:00:00.079) 0:02:08.543 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Saturday 11 January 2025 11:30:43 -0500 (0:00:00.090) 0:02:08.634 ****** ok: [managed-node2] => { "changed": false, "stat": { "atime": 1736612945.6352425, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1736612940.9742577, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 9160785, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "1643853349", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.441) 0:02:09.075 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.079) 0:02:09.154 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.083) 0:02:09.238 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.070) 0:02:09.309 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.069) 0:02:09.378 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.069) 0:02:09.448 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.070) 0:02:09.519 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.051) 0:02:09.570 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:115 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.045) 0:02:09.615 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:124 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.117) 0:02:09.733 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.072) 0:02:09.806 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.039) 0:02:09.845 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:127 Saturday 11 January 2025 11:30:44 -0500 (0:00:00.044) 0:02:09.889 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Saturday 11 January 2025 11:30:45 -0500 (0:00:00.127) 0:02:10.017 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Saturday 11 January 2025 11:30:45 -0500 (0:00:00.070) 0:02:10.087 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:130 Saturday 11 January 2025 11:30:45 -0500 (0:00:00.066) 0:02:10.154 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Saturday 11 January 2025 11:30:45 -0500 (0:00:00.155) 0:02:10.309 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Saturday 11 January 2025 11:30:45 -0500 (0:00:00.071) 0:02:10.380 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:133 Saturday 11 January 2025 11:30:45 -0500 (0:00:00.155) 0:02:10.536 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Saturday 11 January 2025 11:30:45 -0500 (0:00:00.136) 0:02:10.673 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Saturday 11 January 2025 11:30:45 -0500 (0:00:00.071) 0:02:10.744 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Saturday 11 January 2025 11:30:45 -0500 (0:00:00.069) 0:02:10.814 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Saturday 11 January 2025 11:30:45 -0500 (0:00:00.061) 0:02:10.876 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:139 Saturday 11 January 2025 11:30:45 -0500 (0:00:00.065) 0:02:10.941 ****** included: fedora.linux_system_roles.firewall for managed-node2 TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Saturday 11 January 2025 11:30:46 -0500 (0:00:00.204) 0:02:11.146 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed-node2 TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Saturday 11 January 2025 11:30:46 -0500 (0:00:00.133) 0:02:11.279 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Saturday 11 January 2025 11:30:46 -0500 (0:00:00.083) 0:02:11.362 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Saturday 11 January 2025 11:30:46 -0500 (0:00:00.138) 0:02:11.501 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22 Saturday 11 January 2025 11:30:46 -0500 (0:00:00.057) 0:02:11.559 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27 Saturday 11 January 2025 11:30:46 -0500 (0:00:00.049) 0:02:11.609 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __firewall_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 Saturday 11 January 2025 11:30:46 -0500 (0:00:00.043) 0:02:11.653 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: firewalld TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43 Saturday 11 January 2025 11:30:47 -0500 (0:00:00.752) 0:02:12.405 ****** skipping: [managed-node2] => { "false_condition": "__firewall_is_transactional | d(false)" } TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48 Saturday 11 January 2025 11:30:47 -0500 (0:00:00.039) 0:02:12.444 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53 Saturday 11 January 2025 11:30:47 -0500 (0:00:00.041) 0:02:12.485 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Saturday 11 January 2025 11:30:47 -0500 (0:00:00.054) 0:02:12.540 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Saturday 11 January 2025 11:30:47 -0500 (0:00:00.066) 0:02:12.606 ****** skipping: [managed-node2] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [managed-node2] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Saturday 11 January 2025 11:30:47 -0500 (0:00:00.081) 0:02:12.687 ****** ok: [managed-node2] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sat 2025-01-11 11:29:10 EST", "ActiveEnterTimestampMonotonic": "271605433", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus-broker.service dbus.socket sysinit.target polkit.service system.slice basic.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sat 2025-01-11 11:29:10 EST", "AssertTimestampMonotonic": "271333792", "Before": "multi-user.target shutdown.target network-pre.target", "BindLogSockets": "no", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "457719000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanLiveMount": "yes", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2025-01-11 11:29:10 EST", "ConditionTimestampMonotonic": "271333789", "ConfigurationDirectoryMode": "0755", "Conflicts": "iptables.service ipset.service ip6tables.service ebtables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4516", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DebugInvocation": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698229248", "EffectiveMemoryMax": "3698229248", "EffectiveTasksMax": "22365", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainHandoffTimestampMonotonic": "271363582", "ExecMainPID": "10862", "ExecMainStartTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainStartTimestampMonotonic": "271336631", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2025-01-11 11:29:10 EST", "InactiveExitTimestampMonotonic": "271337679", "InvocationID": "f2f268445be048bc90d6f558b31f85a8", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13978", "LimitNPROCSoft": "13978", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13978", "LimitSIGPENDINGSoft": "13978", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LiveMountResult": "success", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "10862", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureDurationUSec": "[not set]", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3241369600", "MemoryCurrent": "33021952", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "35106816", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "0", "MemorySwapMax": "infinity", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivatePIDs": "no", "PrivateTmp": "no", "PrivateTmpEx": "no", "PrivateUsers": "no", "PrivateUsersEx": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectControlGroupsEx": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket sysinit.target system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestamp": "Sat 2025-01-11 11:29:10 EST", "StateChangeTimestampMonotonic": "271605433", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22365", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Saturday 11 January 2025 11:30:48 -0500 (0:00:00.554) 0:02:13.242 ****** ok: [managed-node2] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sat 2025-01-11 11:29:10 EST", "ActiveEnterTimestampMonotonic": "271605433", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus-broker.service dbus.socket sysinit.target polkit.service system.slice basic.target", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sat 2025-01-11 11:29:10 EST", "AssertTimestampMonotonic": "271333792", "Before": "multi-user.target shutdown.target network-pre.target", "BindLogSockets": "no", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "457719000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanLiveMount": "yes", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2025-01-11 11:29:10 EST", "ConditionTimestampMonotonic": "271333789", "ConfigurationDirectoryMode": "0755", "Conflicts": "iptables.service ipset.service ip6tables.service ebtables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4516", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DebugInvocation": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698229248", "EffectiveMemoryMax": "3698229248", "EffectiveTasksMax": "22365", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainHandoffTimestampMonotonic": "271363582", "ExecMainPID": "10862", "ExecMainStartTimestamp": "Sat 2025-01-11 11:29:10 EST", "ExecMainStartTimestampMonotonic": "271336631", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[Sat 2025-01-11 11:29:10 EST] ; stop_time=[n/a] ; pid=10862 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2025-01-11 11:29:10 EST", "InactiveExitTimestampMonotonic": "271337679", "InvocationID": "f2f268445be048bc90d6f558b31f85a8", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13978", "LimitNPROCSoft": "13978", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13978", "LimitSIGPENDINGSoft": "13978", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LiveMountResult": "success", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "10862", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureDurationUSec": "[not set]", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3233042432", "MemoryCurrent": "33021952", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "35106816", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "0", "MemorySwapMax": "infinity", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivatePIDs": "no", "PrivateTmp": "no", "PrivateTmpEx": "no", "PrivateUsers": "no", "PrivateUsersEx": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectControlGroupsEx": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket sysinit.target system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestamp": "Sat 2025-01-11 11:29:10 EST", "StateChangeTimestampMonotonic": "271605433", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22365", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Saturday 11 January 2025 11:30:48 -0500 (0:00:00.696) 0:02:13.938 ****** ok: [managed-node2] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3.12", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Saturday 11 January 2025 11:30:49 -0500 (0:00:00.089) 0:02:14.028 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Saturday 11 January 2025 11:30:49 -0500 (0:00:00.063) 0:02:14.092 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Saturday 11 January 2025 11:30:49 -0500 (0:00:00.063) 0:02:14.155 ****** ok: [managed-node2] => (item={'port': '15001-15003/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "15001-15003/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Saturday 11 January 2025 11:30:49 -0500 (0:00:00.544) 0:02:14.700 ****** skipping: [managed-node2] => (item={'port': '15001-15003/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "'detailed' in fw[0]", "item": { "port": "15001-15003/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed-node2] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Saturday 11 January 2025 11:30:49 -0500 (0:00:00.058) 0:02:14.758 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "'detailed' in fw[0]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Saturday 11 January 2025 11:30:49 -0500 (0:00:00.050) 0:02:14.809 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Saturday 11 January 2025 11:30:49 -0500 (0:00:00.039) 0:02:14.848 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Saturday 11 January 2025 11:30:49 -0500 (0:00:00.039) 0:02:14.888 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Saturday 11 January 2025 11:30:49 -0500 (0:00:00.036) 0:02:14.924 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Saturday 11 January 2025 11:30:49 -0500 (0:00:00.034) 0:02:14.959 ****** skipping: [managed-node2] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:146 Saturday 11 January 2025 11:30:50 -0500 (0:00:00.153) 0:02:15.113 ****** redirecting (type: modules) ansible.builtin.selinux to ansible.posix.selinux redirecting (type: modules) ansible.builtin.selinux to ansible.posix.selinux redirecting (type: modules) ansible.builtin.seboolean to ansible.posix.seboolean included: fedora.linux_system_roles.selinux for managed-node2 TASK [fedora.linux_system_roles.selinux : Set ansible_facts required by role and install packages] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:2 Saturday 11 January 2025 11:30:50 -0500 (0:00:00.215) 0:02:15.328 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml for managed-node2 TASK [fedora.linux_system_roles.selinux : Ensure ansible_facts used by role] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml:2 Saturday 11 January 2025 11:30:50 -0500 (0:00:00.118) 0:02:15.446 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Ensure SELinux packages] ************* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/set_facts_packages.yml:7 Saturday 11 January 2025 11:30:50 -0500 (0:00:00.123) 0:02:15.570 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml for managed-node2 TASK [fedora.linux_system_roles.selinux : Check if system is ostree] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:5 Saturday 11 January 2025 11:30:50 -0500 (0:00:00.081) 0:02:15.651 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set flag to indicate system is ostree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:10 Saturday 11 January 2025 11:30:50 -0500 (0:00:00.051) 0:02:15.703 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:17 Saturday 11 January 2025 11:30:50 -0500 (0:00:00.039) 0:02:15.742 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set flag if transactional-update exists] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:22 Saturday 11 January 2025 11:30:50 -0500 (0:00:00.041) 0:02:15.784 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __selinux_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux python2 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:26 Saturday 11 January 2025 11:30:50 -0500 (0:00:00.041) 0:02:15.825 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_python_version is version('3', '<')", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux python3 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:35 Saturday 11 January 2025 11:30:50 -0500 (0:00:00.099) 0:02:15.925 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: python3-libselinux python3-policycoreutils TASK [fedora.linux_system_roles.selinux : Install SELinux python3 tools] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:46 Saturday 11 January 2025 11:30:51 -0500 (0:00:00.762) 0:02:16.687 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_os_family == \"Suse\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Install SELinux tool semanage] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:58 Saturday 11 January 2025 11:30:51 -0500 (0:00:00.049) 0:02:16.737 ****** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: policycoreutils-python-utils TASK [fedora.linux_system_roles.selinux : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:72 Saturday 11 January 2025 11:30:52 -0500 (0:00:00.785) 0:02:17.522 ****** skipping: [managed-node2] => { "false_condition": "__selinux_is_transactional | d(false)" } TASK [fedora.linux_system_roles.selinux : Reboot transactional update systems] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:77 Saturday 11 January 2025 11:30:52 -0500 (0:00:00.081) 0:02:17.604 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Fail if reboot is needed and not set] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:82 Saturday 11 January 2025 11:30:52 -0500 (0:00:00.092) 0:02:17.696 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__selinux_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Refresh facts] *********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:89 Saturday 11 January 2025 11:30:52 -0500 (0:00:00.072) 0:02:17.769 ****** ok: [managed-node2] TASK [fedora.linux_system_roles.selinux : Set permanent SELinux state if enabled] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:5 Saturday 11 January 2025 11:30:53 -0500 (0:00:00.911) 0:02:18.680 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_selinux.status == \"enabled\" and (selinux_state or selinux_policy)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set permanent SELinux state if disabled] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:13 Saturday 11 January 2025 11:30:53 -0500 (0:00:00.055) 0:02:18.736 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "ansible_selinux.status == \"disabled\" and selinux_state", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set selinux_reboot_required] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:21 Saturday 11 January 2025 11:30:53 -0500 (0:00:00.039) 0:02:18.776 ****** ok: [managed-node2] => { "ansible_facts": { "selinux_reboot_required": false }, "changed": false } TASK [fedora.linux_system_roles.selinux : Fail if reboot is required] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:25 Saturday 11 January 2025 11:30:53 -0500 (0:00:00.047) 0:02:18.824 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_reboot_required", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Warn if SELinux is disabled] ********* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:30 Saturday 11 January 2025 11:30:53 -0500 (0:00:00.041) 0:02:18.866 ****** skipping: [managed-node2] => { "false_condition": "ansible_selinux.status == \"disabled\"" } TASK [fedora.linux_system_roles.selinux : Drop all local modifications] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:35 Saturday 11 January 2025 11:30:54 -0500 (0:00:00.151) 0:02:19.017 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_all_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux boolean local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:43 Saturday 11 January 2025 11:30:54 -0500 (0:00:00.065) 0:02:19.083 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_booleans_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux file context local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:48 Saturday 11 January 2025 11:30:54 -0500 (0:00:00.066) 0:02:19.149 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_fcontexts_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux port local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:53 Saturday 11 January 2025 11:30:54 -0500 (0:00:00.069) 0:02:19.219 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_ports_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Purge all SELinux login local modifications] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:58 Saturday 11 January 2025 11:30:54 -0500 (0:00:00.067) 0:02:19.286 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_logins_purge | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Set SELinux booleans] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:63 Saturday 11 January 2025 11:30:54 -0500 (0:00:00.061) 0:02:19.348 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Set SELinux file contexts] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:74 Saturday 11 January 2025 11:30:54 -0500 (0:00:00.060) 0:02:19.408 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Set an SELinux label on a port] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:87 Saturday 11 January 2025 11:30:54 -0500 (0:00:00.061) 0:02:19.470 ****** ok: [managed-node2] => (item={'ports': '15001-15003', 'setype': 'http_port_t'}) => { "__selinux_item": { "ports": "15001-15003", "setype": "http_port_t" }, "ansible_loop_var": "__selinux_item", "changed": false, "ports": [ "15001-15003" ], "proto": "tcp", "setype": "http_port_t", "state": "present" } TASK [fedora.linux_system_roles.selinux : Set linux user to SELinux user mapping] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:99 Saturday 11 January 2025 11:30:55 -0500 (0:00:00.677) 0:02:20.148 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Get SELinux modules facts] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:112 Saturday 11 January 2025 11:30:55 -0500 (0:00:00.050) 0:02:20.198 ****** ok: [managed-node2] => { "ansible_facts": { "selinux_checksums": true, "selinux_installed_modules": { "abrt": { "100": { "checksum": "sha256:5fdd1acb3c17ce67f51edc9160653adada9aade378218aed1b38acc0b4d4157a", "enabled": 1 } }, "accountsd": { "100": { "checksum": "sha256:50197ed40bbac67de44c12d745bc4a1259c8ae3f3828f5b4cfcd02b187ee1f47", "enabled": 1 } }, "acct": { "100": { "checksum": "sha256:14149edb1593dad6e684a672c57ca35f2ff51aee24ef503171c28671dc1bff13", "enabled": 1 } }, "afs": { "100": { "checksum": "sha256:769aaf25a8699233ceeda87ac858ebc91f094ea98b303c54e2c810760acd9756", "enabled": 1 } }, "afterburn": { "100": { "checksum": "sha256:9b9c00f12dea94989d2f5c592fa8ba4cff49b1ae7c9636fc38baa7957750e80f", "enabled": 1 } }, "aiccu": { "100": { "checksum": "sha256:d7ed1435298abcb42678af33c66ba2bac4c02723601c5018ecef64ca4c4e8f0b", "enabled": 1 } }, "aide": { "100": { "checksum": "sha256:95c6046b0d690df4792fe4418de5f7753279b72c78c5f1dab5397aba8bdf7152", "enabled": 1 } }, "alsa": { "100": { "checksum": "sha256:c0be471cd7182d1126cd6d19b276742a6ace62429425a81d42d8af4c24bcdbb4", "enabled": 1 } }, "amanda": { "100": { "checksum": "sha256:2ea0f86264cad0fa7f1eba0e6238526e92fadc65fd9754f62bc6d59b2da9086c", "enabled": 1 } }, "amtu": { "100": { "checksum": "sha256:1c77cc13007a0024e3121f064aad7dafce5fd57d7f71140f0af6b5af10805a8b", "enabled": 1 } }, "anaconda": { "100": { "checksum": "sha256:79386dbc8f95842f819ddbb01ce5824752d78bb4992783288d79db391a4ec3b5", "enabled": 1 } }, "antivirus": { "100": { "checksum": "sha256:48aafc3e308b4db80797ec75f52d3cc1b8216315d28108a10a03e47a0209adda", "enabled": 1 } }, "apache": { "100": { "checksum": "sha256:30a1e157e56f57a82557524173635e911f796fa8733f98ad190fc854aba893d8", "enabled": 1 } }, "apcupsd": { "100": { "checksum": "sha256:46a2e84d4dd0064f0b368787ed9c494ace4e337d5d80bcc6db77f66dede34375", "enabled": 1 } }, "apm": { "100": { "checksum": "sha256:4f5ff9b6ad61bc6a311bc71768e7ead7fe273989b7cf2fa58b1a13221e8d82f4", "enabled": 1 } }, "application": { "100": { "checksum": "sha256:ee11bb52bc37177920d348c4b920738a1b64413f516c66e18fa5a453a6d81f5e", "enabled": 1 } }, "arpwatch": { "100": { "checksum": "sha256:11a3c765598ae81cda7af561ada4795fb53a9ae9cfcb02af806b08cfc7f00b0e", "enabled": 1 } }, "asterisk": { "100": { "checksum": "sha256:ce887eb790f85810eb5eec0ac96a4e8028d633cf16fc117db4f733e6040c11c2", "enabled": 1 } }, "auditadm": { "100": { "checksum": "sha256:201655fdd63b544aa448c7b46c8b7ec8aa726ef3b55e9ef928ff3502c128973b", "enabled": 1 } }, "authlogin": { "100": { "checksum": "sha256:f1937fa4d112f634f69b9f0104872ce507c56928951a961c26ddf385a12be54e", "enabled": 1 } }, "automount": { "100": { "checksum": "sha256:5d308d3654fc5b0cea77156c27b98bbe2e370b843a920d75e1d616472caee55f", "enabled": 1 } }, "avahi": { "100": { "checksum": "sha256:ab1c62da7d38c1870efff54480b9a43ac2683eb0ebd77925a6fdea0fd47e207c", "enabled": 1 } }, "awstats": { "100": { "checksum": "sha256:f4a83008d582e61321a1d5462807c402d6cd3034d272dab00dd788122f245114", "enabled": 1 } }, "bacula": { "100": { "checksum": "sha256:04e3c063e146bd0f96587e9f49cdb4020e83bf69ba77250d61d2c9fcea53476b", "enabled": 1 } }, "base": { "100": { "checksum": "sha256:106220779e72010bf08e45ebec252a2e47257ca1134c07dbcd68a0d37e790931", "enabled": 1 } }, "bcfg2": { "100": { "checksum": "sha256:860a81a71f68a66e7dc0cee62e02f5da8d2dbb83724bfde9abc24b56b06ec914", "enabled": 1 } }, "bind": { "100": { "checksum": "sha256:ab2767aacd75df866fc6080d7db8dfefbd013a4ad578412ccfa6dece0d0b8ea7", "enabled": 1 } }, "bitlbee": { "100": { "checksum": "sha256:6ae2a654612810131bc874411156b1ef5d784143484e2a867ef413b6ab0f8750", "enabled": 1 } }, "blkmapd": { "100": { "checksum": "sha256:a6ac037c876fcea13d9457717138ec1c17913a0a303469219aa610351225c30c", "enabled": 1 } }, "blueman": { "100": { "checksum": "sha256:10372514dd684076a8cc4c93211d6f6732dcff0fa6aceed02dce868b12aaee82", "enabled": 1 } }, "bluetooth": { "100": { "checksum": "sha256:d8e9f771f666c02cacd7881f83cebd64ca88d79382fddca9e608d75df65099a2", "enabled": 1 } }, "boinc": { "100": { "checksum": "sha256:4ede0a1fb4501325ad8d0ff576f2f20bb0c3e994d2142372449de840f10e8b4b", "enabled": 1 } }, "boltd": { "100": { "checksum": "sha256:63a74e88f61411cfd03aa6e719867d1ae3238b99f3adb2f91cc0973e8b10e174", "enabled": 1 } }, "boothd": { "100": { "checksum": "sha256:4df95e4f92b1da81af245d74fd4481aad048554cad1d31f87985e3c68814ac8e", "enabled": 1 } }, "bootloader": { "100": { "checksum": "sha256:c4b351c2ad386a4c22a477788aeba9ae713cddaae623d733f1957d5d521c717d", "enabled": 1 } }, "bootupd": { "100": { "checksum": "sha256:fd688c68b40c042dd04fe9cb65c7a270353a2f7c677a5e73864b34ee6118f94d", "enabled": 1 } }, "brctl": { "100": { "checksum": "sha256:3a931c1c282cee1b53f6aee3730fef0e498fba3c5005218b4c7a923a7ce226c2", "enabled": 1 } }, "brltty": { "100": { "checksum": "sha256:0471f4ad4a09025a1a23ad398d869d88b90172a869800cf06660c886033c7409", "enabled": 1 } }, "bugzilla": { "100": { "checksum": "sha256:480fe3690a301d2359b747a03b7f65fd881063f6f7d877c55a16e6a51f324df6", "enabled": 1 } }, "cachefilesd": { "100": { "checksum": "sha256:04eafc0b0c7783758bba1cb55893416c204933bf0fdef964ce23c6f17c626c65", "enabled": 1 } }, "calamaris": { "100": { "checksum": "sha256:3b369138db2945167e6b10503d99902411fd5213b91dae978afe344b174240f7", "enabled": 1 } }, "callweaver": { "100": { "checksum": "sha256:0d66c1686efea20f479f6dd3057f5bd5550709d34e3a30745900abf363728bf1", "enabled": 1 } }, "canna": { "100": { "checksum": "sha256:5677c14448ec90dd564b7fe2c19146faa264a6cc80da75dba8f51c1e17af73d2", "enabled": 1 } }, "ccs": { "100": { "checksum": "sha256:ab0abd34c688230d836833a4f46228f3873ccbcfd3fd351d1306c203368ab63b", "enabled": 1 } }, "cdrecord": { "100": { "checksum": "sha256:c258276b231b731865f4237c8477010ea00b71e0e51501d7a206344381f625b1", "enabled": 1 } }, "certmaster": { "100": { "checksum": "sha256:5d46915b4fa32deccdf587b5970cde875d7b43643c09ec42a156193f9592c8ab", "enabled": 1 } }, "certmonger": { "100": { "checksum": "sha256:8746abfbb345c4bfd16b1c27e127a281e72388536aa9c7545ed308a266aceeb5", "enabled": 1 } }, "certwatch": { "100": { "checksum": "sha256:8d8299ad27c884a1657ff8ba52413f24867605e5e31c55a70659b3b0db4309d5", "enabled": 1 } }, "cfengine": { "100": { "checksum": "sha256:c3e0da50f4e8cd55d3a139afdcfa0fc5479b0e4afe78df4a1eab9400a71b64e8", "enabled": 1 } }, "cgroup": { "100": { "checksum": "sha256:e681dc6218ef1387435a6c4e95ed13e45d5edf2089991beb9bc8df41021cd865", "enabled": 1 } }, "chrome": { "100": { "checksum": "sha256:d64d5c77ef6d19be3aecf839fc444fab44a5717fa879c6d1bf6eb7564f556835", "enabled": 1 } }, "chronyd": { "100": { "checksum": "sha256:ea2b23dec628aecd68d6b8bc7d88d662e6bc0e72330f2a87e436f66a38b28ed4", "enabled": 1 } }, "cifsutils": { "100": { "checksum": "sha256:db66a0e17ecf7229eff4ffe84885bbff334085e3e20c282dd7e1eab97545e5e7", "enabled": 1 } }, "cinder": { "100": { "checksum": "sha256:95351f4464622aed05297db68b883d47165ce7db322134048f5a99479c734da3", "enabled": 1 } }, "cipe": { "100": { "checksum": "sha256:25016148cbfa7480cee4b614572bf7a364f6e45a5aca844bb9a87bba217b93a5", "enabled": 1 } }, "clock": { "100": { "checksum": "sha256:5f8b31db8708926fe3de99aadf34b5114184bff20956c7203c6329433bc24dc7", "enabled": 1 } }, "clogd": { "100": { "checksum": "sha256:b6bdd15aeae4574bffcd8a3c862261eb9e21ae5451035cc470ee9eba6cf49cd4", "enabled": 1 } }, "cloudform": { "100": { "checksum": "sha256:8d891dca8241582619e85b4d6f5aa273bd87d3bb7ff04199a38c91f47df12403", "enabled": 1 } }, "cmirrord": { "100": { "checksum": "sha256:e9ccb95f73f750fa464aa4fb5b76e5bad42050e27a128795826272ace5f70760", "enabled": 1 } }, "cobbler": { "100": { "checksum": "sha256:e090b7686819373bcad8a0cdfcf17dd453f7873dfe757742b4bee5c2be7f58d5", "enabled": 1 } }, "collectd": { "100": { "checksum": "sha256:96c655a73312da543bc2aa652b14079cb6fdac677618325d9f1ad83996dbeee1", "enabled": 1 } }, "colord": { "100": { "checksum": "sha256:180a83a6a6059ef9044f9a171ec71a5934a5a19da0f9ce5451c3ba1bbfb75343", "enabled": 1 } }, "comsat": { "100": { "checksum": "sha256:d1e10fcd329ec130a7346374dd6cf339eb38e2c37bc05f371c0381cc7e022d2e", "enabled": 1 } }, "condor": { "100": { "checksum": "sha256:f27781ec40ba6625c3854cf0e6c311a345d5619a3b07c7105635dc60a9369a6c", "enabled": 1 } }, "conman": { "100": { "checksum": "sha256:7984cbaf5572913966bae94db45941e7e15fc195f70c0bd30cbd6b8a24998687", "enabled": 1 } }, "conntrackd": { "100": { "checksum": "sha256:0607c19ac5763806c4420869caf11320ecad25b9ba9e10307105c75e9a26f6be", "enabled": 1 } }, "consolekit": { "100": { "checksum": "sha256:e6e39c814a93fa8973beffb8f891030f7b8094d9987e221d140c1a1fc79a196b", "enabled": 1 } }, "coreos_installer": { "100": { "checksum": "sha256:0ca9390a1bd0c5caa0958c03c1f665a129836f50e3ee11d54fa21800fab9aa08", "enabled": 1 } }, "couchdb": { "100": { "checksum": "sha256:e7ffb1afd19042410ffc4af01ee0fe31ff9f434d23451dd3bd01a7052be716b2", "enabled": 1 } }, "courier": { "100": { "checksum": "sha256:113abb8c428ab21653d9c2f9aef35354470381234ccb493e882ef22f72c2ede4", "enabled": 1 } }, "cpucontrol": { "100": { "checksum": "sha256:36939ecc2515a7980f3be7a2d9518018b3502aa526a446b04552f43a671518dc", "enabled": 1 } }, "cpufreqselector": { "100": { "checksum": "sha256:43b7f9d92ade2e28e26c7f9736d94d91b9a20812118f466b39aa4dc7a02e0959", "enabled": 1 } }, "cpuplug": { "100": { "checksum": "sha256:97a1c9de6df462cab0645b58e16072042c4cab7704c10209950684673a049b8d", "enabled": 1 } }, "cron": { "100": { "checksum": "sha256:0e02e661cc66e751e27d9ece16db633d5ab2036dab470b9dd4a1ef0725bf1534", "enabled": 1 } }, "ctdb": { "100": { "checksum": "sha256:7a7a3998f8a29178005386630556039bba9812f206b5f6782bdcfc0c22f66cbc", "enabled": 1 } }, "cups": { "100": { "checksum": "sha256:90605693f95fba47762dba603d99f0603ccea7238937bd0ed4ca8dc680fd7382", "enabled": 1 } }, "cvs": { "100": { "checksum": "sha256:05274fef587a5ac9def4bce8be2540bc893dd2c7b0052ad9f4e5de0c6da55718", "enabled": 1 } }, "cyphesis": { "100": { "checksum": "sha256:635739376db164b36a935a3bb0dbe8b5644b0fca8d3791b0efde3da36372a4aa", "enabled": 1 } }, "cyrus": { "100": { "checksum": "sha256:26567d47104e4634ab34aa39e0a3ea5f4ed85670b6c66b63e86c47c809080344", "enabled": 1 } }, "daemontools": { "100": { "checksum": "sha256:429c5e1b65232dbfb74741ba1514315288e2704c8a105759c31b85e3742200e1", "enabled": 1 } }, "dbadm": { "100": { "checksum": "sha256:24d6e3dfb505a297328108979f63cbad1b2a0f241c2fb2e84f93dadc086f1273", "enabled": 1 } }, "dbskk": { "100": { "checksum": "sha256:0bf031897570904400ae864c9d12f8880047d1c0368ebb8edc7fd7232f2eecab", "enabled": 1 } }, "dbus": { "100": { "checksum": "sha256:c0b2fa3d02d462ddbf8269485861a3e6e873cd5f05a174a99b121095cbd31f97", "enabled": 1 } }, "dcc": { "100": { "checksum": "sha256:f237213b8df023a69261f26318e3572f28a2c03bccf8d37505dabeca64aa9fe4", "enabled": 1 } }, "ddclient": { "100": { "checksum": "sha256:4159c1afa1a0e1977328780489eb61c30f61823b389f4cd0d79f25a7b27d5857", "enabled": 1 } }, "denyhosts": { "100": { "checksum": "sha256:643ce1a1f7dc2df66e3a1ff57d3dbd9631ea7c9b9b83467d7e4e2f4dc701186b", "enabled": 1 } }, "devicekit": { "100": { "checksum": "sha256:845cf9ac75f705a72df5d2b643b810fe266919369f6840e461f9d3043db97bb9", "enabled": 1 } }, "dhcp": { "100": { "checksum": "sha256:57fe91a9ccded7037429132ca5c28c408881f296eb8e603041b61f82c673c7c6", "enabled": 1 } }, "dictd": { "100": { "checksum": "sha256:246922587bf734f34d26ddb793ab9c649460b3cea63acfca54dc8a989e948d97", "enabled": 1 } }, "dirsrv": { "100": { "checksum": "sha256:9c67bc04c2d69254c0421c864b6f309c4c213bf415759b93a11a21bafa55b7f2", "enabled": 1 } }, "distcc": { "100": { "checksum": "sha256:a8179c2e64684b994ca8802badba7dbe0df42564f400307eeca0274f442cc0c4", "enabled": 1 } }, "dmesg": { "100": { "checksum": "sha256:977de79bba759fc421b791cc5496b0479b34077259eb0b9aab7049cce9e3d537", "enabled": 1 } }, "dmidecode": { "100": { "checksum": "sha256:dbe77a5d3671052e5b1dff63016c24e6e576803c82edbf2a913147a24f893124", "enabled": 1 } }, "dnsmasq": { "100": { "checksum": "sha256:7ad2d3372989319aaef08ac68aff266ae10dd262728284d4531851e893fe9edf", "enabled": 1 } }, "dnssec": { "100": { "checksum": "sha256:70ff660118b3bb9c8e5107d566adf6c2ad0b4af94d6757852c68a00eb19f6e56", "enabled": 1 } }, "dovecot": { "100": { "checksum": "sha256:1057c6ca706e9266c2e147865fb58e634439b7ec3e6c1faba0185401e2d4e985", "enabled": 1 } }, "drbd": { "100": { "checksum": "sha256:225633f600a5c31fcb19a9b721677833e32b88e57ef8cfaa3fa88b25bfc87c87", "enabled": 1 } }, "dspam": { "100": { "checksum": "sha256:503e9b509a0f764faea4a2415311aa1b43bfd8b78bbcc8090f6bc81c200beea8", "enabled": 1 } }, "entropyd": { "100": { "checksum": "sha256:e2aa3a7bc12f543b79842c2504cb1b53b73c55c78a42e24fbead6ab2deb120a4", "enabled": 1 } }, "exim": { "100": { "checksum": "sha256:b097146234c292a604496276a05ae2cbb09bbfdbf0ac9d7948b42afddd953347", "enabled": 1 } }, "fail2ban": { "100": { "checksum": "sha256:c51270c9abffd1d173532bbce08b28480f8e261d099370268b5364c343726140", "enabled": 1 } }, "fcoe": { "100": { "checksum": "sha256:a8a2c86fcc37532b9bf516e070ff53c9e675cb5e9f154966dde6a9b5fc6d7bc5", "enabled": 1 } }, "fdo": { "100": { "checksum": "sha256:7ecdacdeaddd502ea66686e0eaa24fa19fef7202f0092a7174db2aa86eddae8e", "enabled": 1 } }, "fedoratp": { "100": { "checksum": "sha256:85ee3ad2c50bf2b0ba1fb4fe0c57850f9ffd3bb5c11d11a5b5e159ac65783c18", "enabled": 1 } }, "fetchmail": { "100": { "checksum": "sha256:56ef87427ba73e1981192f595dd2c169b322cbed91da8de818e995434e07c757", "enabled": 1 } }, "finger": { "100": { "checksum": "sha256:9aa41b846e607a383c12d0bb4a9695058f8c9b1ed4b26bb59a96855015d9ab35", "enabled": 1 } }, "firewalld": { "100": { "checksum": "sha256:9e85ce564ebec8cae461d093727c4e339a32b10ce3efa569b5c2d53c27baae59", "enabled": 1 } }, "firewallgui": { "100": { "checksum": "sha256:433e16150ab0650ca51df171c1fbe1406cfe15dfa077bade2e48a7d8514bb0ba", "enabled": 1 } }, "firstboot": { "100": { "checksum": "sha256:a16615a4aea8b12fce51f6d7715627da3c3f6c6099db8c3087e29c920d81f861", "enabled": 1 } }, "fprintd": { "100": { "checksum": "sha256:1cb012beb208aebe26b53919c54e8e902210feb7b26d52b05cbe760d273b0645", "enabled": 1 } }, "freeipmi": { "100": { "checksum": "sha256:f62cf10a14cb8de0f158fed8cc9400a8769be1ffb2f3dcdcd0e920de20d68007", "enabled": 1 } }, "freqset": { "100": { "checksum": "sha256:5c96b7146bde0acec28d73afc25b0a35fbfffe2ccd69529717eca6927dc558a7", "enabled": 1 } }, "fstools": { "100": { "checksum": "sha256:0fdc4babffbc8fa18d9445b74b14fc08b77fd6e371c96dc1f5283d6ff7c5f546", "enabled": 1 } }, "ftp": { "100": { "checksum": "sha256:dae3209382de68fdb49a72792108ec3a55c0e50abdfd38c5fae300da4c561c47", "enabled": 1 } }, "fwupd": { "100": { "checksum": "sha256:4f1c84d8ccd64b0af49edc9247ac92f49545fe4dfb24906acc133ebd7ba78bb9", "enabled": 1 } }, "games": { "100": { "checksum": "sha256:7d9b09309f694125d4ba8e1f5683baedfc2c935f0242be94531b43d842509ae9", "enabled": 1 } }, "gdomap": { "100": { "checksum": "sha256:bec0b003cf82cc2fe7e6f196bd82b929e87a827fe8d46bca5b1df2613a409d6b", "enabled": 1 } }, "geoclue": { "100": { "checksum": "sha256:3e8a90189684079f66e782c8e001c435ada771766dd1fa9ae33a977405f6ba99", "enabled": 1 } }, "getty": { "100": { "checksum": "sha256:e5ecf805b71dfe0f039f8eeefa730893c5417fe53b0e0eeba55f99670b4c4584", "enabled": 1 } }, "git": { "100": { "checksum": "sha256:69850d2342f0fcb96a719d39a3698d547c105880e6834f232a4d845a18a046f3", "enabled": 1 } }, "gitosis": { "100": { "checksum": "sha256:24935d3495375c7f6e0a0ddc091cf1f3f276ee155d07760bd7ba0c3ac115f6e8", "enabled": 1 } }, "glance": { "100": { "checksum": "sha256:fe00543de47f940ba5bf1ab84b7dc7f6a680053ef148329f2a9ec827eb47c61c", "enabled": 1 } }, "glusterd": { "100": { "checksum": "sha256:4d2befdcd924820622a0c6653772a9f5ba96aad9a3bd7b4d86daab96b233c9ea", "enabled": 1 } }, "gnome": { "100": { "checksum": "sha256:a5e2adeeda1a5b359025d956cbf20eb3773341207c72cdf7bf5001deeedf4712", "enabled": 1 } }, "gnome_remote_desktop": { "100": { "checksum": "sha256:d926eb26b34a6253b249bc1cea7696dae87887739ef77cd8cb92bdf7f6d1f371", "enabled": 1 } }, "gpg": { "100": { "checksum": "sha256:a9180cca4098d53514cd5f701f4afd40420a3664e8ffa86797ee3a8fadd9a84a", "enabled": 1 } }, "gpm": { "100": { "checksum": "sha256:2b5e83f7513a2470766cbaad46e28604ed8fcd6b640dc0350d90731a807f56ad", "enabled": 1 } }, "gpsd": { "100": { "checksum": "sha256:b64db8adeb9e8602211a0924c1b4b917d0f701df86984e73412344c9b3c2f5d9", "enabled": 1 } }, "gssproxy": { "100": { "checksum": "sha256:a43f376f2b1605cf0d45884b818db045eb37b4f037ae7af36ec5bf993676f4e5", "enabled": 1 } }, "guest": { "100": { "checksum": "sha256:262ccc7c3b237e7ebc31631d83baf2c6cbf8438cce02ce82f008fd6a3ff81a43", "enabled": 1 } }, "hddtemp": { "100": { "checksum": "sha256:fc86a156d59e31b02ce40cbda3c24b0c0baae2ed6f0cabf84cb9a3248ccd236e", "enabled": 1 } }, "hostapd": { "100": { "checksum": "sha256:ba13f04311a741f26ac34cd39688e2f4e75a546358c63c57a2d6a916dfc18315", "enabled": 1 } }, "hostname": { "100": { "checksum": "sha256:7366a48a1774b579953d8b8ae05f9ebe07e5d27984c995afe27509eacf1d79d6", "enabled": 1 } }, "hsqldb": { "100": { "checksum": "sha256:a0c31e8462c7fbbc6579270e5e35923deff56790f565157cb9c32c4eb3d6958b", "enabled": 1 } }, "hwloc": { "100": { "checksum": "sha256:fde041b834e5939945fc0d91e45940745254e128d96625ca5ab4d6e7a4379e27", "enabled": 1 } }, "hypervkvp": { "100": { "checksum": "sha256:520caa8d38f07a36d096f7ee8eb1b0610f49eafedca1721b340f5cdf21e23944", "enabled": 1 } }, "ibacm": { "100": { "checksum": "sha256:4bc0f9a40691f1abb15ce6cd277d7bab4b18720373a197c8310aa63995bcda0f", "enabled": 1 } }, "ica": { "100": { "checksum": "sha256:5daeb045f8b76fed05c1145d131f77dd556de131f777b7e259d8ec57787e7da6", "enabled": 1 } }, "icecast": { "100": { "checksum": "sha256:281ed7601c5b057802da5ffdc24a7e3308e0bb2591a263186f51e35816dffc12", "enabled": 1 } }, "iiosensorproxy": { "100": { "checksum": "sha256:05dd0967ab5bc2422ac57f2ec8a354181bf7038f0d1aabc2775c7c915bbbeaf3", "enabled": 1 } }, "inetd": { "100": { "checksum": "sha256:52f6df31e38b0c20b9fc264115ce56b4d6d8048dce5b6f250a6ca1b0675a0569", "enabled": 1 } }, "init": { "100": { "checksum": "sha256:7141bbe724e0c7f4b3dd7f335ab21267676b126f0443da9cbfa58f1372deb59c", "enabled": 1 } }, "inn": { "100": { "checksum": "sha256:78ae9628e756da216964c88e987dd91e40ec2803ff209763f628a434c8bf3b90", "enabled": 1 } }, "insights_client": { "100": { "checksum": "sha256:c6e547e12ca21f1f68ba74d667fab1bcbfc8d0461fc3a560e80160f32638ddba", "enabled": 1 } }, "iodine": { "100": { "checksum": "sha256:5c8ef0d8e730a3baca51f4c64b13df56149ef304effdc71fe0fb63f90e5a7296", "enabled": 1 } }, "iotop": { "100": { "checksum": "sha256:114de54fd32be9c32e88348ddb168291c0aaaba3a7937501f49e514585228092", "enabled": 1 } }, "ipmievd": { "100": { "checksum": "sha256:5777d2cc5e62942022a8e914847dc7436264585e41e2ce703c49c735e44add8c", "enabled": 1 } }, "ipsec": { "100": { "checksum": "sha256:ffde25d3dd0fa80b933b30cd457ea900389f84d13d4ea629f54c0be59136d4a6", "enabled": 1 } }, "iptables": { "100": { "checksum": "sha256:4c92a7102c209b0edf571a3ca3b35c597bd427d0d258c6829bfd29358448598b", "enabled": 1 } }, "irc": { "100": { "checksum": "sha256:291a822392f8a93d820aa03613bd0c058cfd1118fd6cca93a12175456eb2cf72", "enabled": 1 } }, "irqbalance": { "100": { "checksum": "sha256:2fc17d69e2cba360a0d9043585a3799f86697bf9189fcde62346b41bcf5bd9d9", "enabled": 1 } }, "iscsi": { "100": { "checksum": "sha256:a91d9d18f0502679892047e7a6023107b77412ae4b3d25f75982f766c792be58", "enabled": 1 } }, "isns": { "100": { "checksum": "sha256:5372fd0c432735b6ec28e0a5cafeebd74688912c7703616ca46b77287385c564", "enabled": 1 } }, "jabber": { "100": { "checksum": "sha256:d1916b00ece1c2b546cc04a7b6695f6459724a6bb7a92be78f399f61a613af86", "enabled": 1 } }, "jetty": { "100": { "checksum": "sha256:c78ba53bd379a65017e29895da717fa8e758494e94113b7f324d73121cb35423", "enabled": 1 } }, "jockey": { "100": { "checksum": "sha256:4cee27b59fe57c672587d220f3061417207a02bc0ff36188148f5c5f3101c1ad", "enabled": 1 } }, "journalctl": { "100": { "checksum": "sha256:27aeef17f22b9e891c4c793870f1124ec69ec7dc8f532bfec5b2a0e7319866cb", "enabled": 1 } }, "kafs": { "100": { "checksum": "sha256:602c3c40b28b653e11110e83c7f63122a75b9ad5cb7605299f6951a485284727", "enabled": 1 } }, "kdump": { "100": { "checksum": "sha256:bac33f3ca081f5f4b58fbbdbd3630aef7fb77c19368ec2cef66494cb8ef7fd90", "enabled": 1 } }, "kdumpgui": { "100": { "checksum": "sha256:7e16a86780b3b3d8a6afc5a09335cd6bf6bb39eec53815af9bbc53844655846c", "enabled": 1 } }, "keepalived": { "100": { "checksum": "sha256:5e76b1d8c27eea054956c86458f44ebc46d803035e2b3c0af67fd5feae26b9f2", "enabled": 1 } }, "kerberos": { "100": { "checksum": "sha256:65a81f3d94068c6e7cd3c1092cfbc080f853c7e02490356f7f16d40a3e8ace49", "enabled": 1 } }, "keyboardd": { "100": { "checksum": "sha256:e267709b6fc5f39679c3156b764ac82facb0b228e5b300d891f4a182662a1a5b", "enabled": 1 } }, "keystone": { "100": { "checksum": "sha256:1a7021842ad7350bc5674a455f4232b831ec8153d14635eba8930e8ade0980b7", "enabled": 1 } }, "keyutils": { "100": { "checksum": "sha256:0dc521ba68e3d21396cf5bce20427e6c0c5dfe635c743ccd4374beadd1a96cc5", "enabled": 1 } }, "kismet": { "100": { "checksum": "sha256:4398910d875745fbd61926116eddc0c7f3f6da3cf8b62fb4149d78734c344df3", "enabled": 1 } }, "kpatch": { "100": { "checksum": "sha256:36ae9280a2f7e86befeaf301f1d47087b9214bc97e79562105fcaa1246512467", "enabled": 1 } }, "ksmtuned": { "100": { "checksum": "sha256:66f6fee140f5f767906ea257b32ede5d214b8da0db4031313eb1fddb9a589d21", "enabled": 1 } }, "ktalk": { "100": { "checksum": "sha256:e783786109e8dac25e8be80e5782ec260343cf0db5b03126b85811a101bd9d0e", "enabled": 1 } }, "l2tp": { "100": { "checksum": "sha256:55f189e95389ac2f9205270f3dd53e25761d8e626989bdfb3c05b19c0086dc30", "enabled": 1 } }, "ldap": { "100": { "checksum": "sha256:cb97aa280c3bb13b6c8804e2a2bf971250db72b77ba5ece99702ab6434c3d14a", "enabled": 1 } }, "libraries": { "100": { "checksum": "sha256:bb45e05979c90858a0ed2ed0ecae78cc4e99b750145947a2940eb0493249d161", "enabled": 1 } }, "likewise": { "100": { "checksum": "sha256:f535c1e16dc83389fd5a2fa4716615c687a931e3a9ad7f65dc8607ef305f7994", "enabled": 1 } }, "linuxptp": { "100": { "checksum": "sha256:0d6f39eec4849cc1f5ca0014c29399b78317216ed0e73dfc524be3194ff1bc08", "enabled": 1 } }, "lircd": { "100": { "checksum": "sha256:cdd2ad3d9fa15b37c76e5bf0d781314f3d50897ef6cf377d70706d8a139639f7", "enabled": 1 } }, "livecd": { "100": { "checksum": "sha256:46a683c9c0693f57e0a687926c0ea32a8a176eaa168d1bf86e66463395b314cf", "enabled": 1 } }, "lldpad": { "100": { "checksum": "sha256:9fa26b1dd3ca8e52c93a1320fb66fceb0c8e83730227023d221a6a426b9dddf0", "enabled": 1 } }, "loadkeys": { "100": { "checksum": "sha256:8e876286cd84dd189d40234f691174bc5e0c214056ff0a9f454400cdafdcd6cd", "enabled": 1 } }, "locallogin": { "100": { "checksum": "sha256:f90a7ca564fcff2da3a9effa5c50e8c2a92de3e4c3831f741bc3f8cfc5be5808", "enabled": 1 } }, "lockdev": { "100": { "checksum": "sha256:d181bc52a0d2af9fe776462be417f9f28920a71f01f9894557e9bf7e128ddb5c", "enabled": 1 } }, "logadm": { "100": { "checksum": "sha256:c573637379d69beb37674b0ce9da5f57e60238a88ddad2faa1c3d456cd7df71a", "enabled": 1 } }, "logging": { "100": { "checksum": "sha256:d3494d1da872d6c128a1ab751a7ec84abb1ec73dccbfae200707935617fb8888", "enabled": 1 } }, "logrotate": { "100": { "checksum": "sha256:97dfef99f123dd0474194e78c4e8760df60d46db61f8460161bbcf4b78f7c530", "enabled": 1 } }, "logwatch": { "100": { "checksum": "sha256:c68409b41b76b32647af8f19d1f8f6ebd43430dc820e55d2b094730950a96419", "enabled": 1 } }, "lpd": { "100": { "checksum": "sha256:24594a8f8d0e864712707e8c4114f6d68dd19824045a9f3d10b8f590fc506858", "enabled": 1 } }, "lsm": { "100": { "checksum": "sha256:7df22b87a981a09999b90bb4ecea9713b595814a385fce77eb67e3c20042aa09", "enabled": 1 } }, "lttng-tools": { "100": { "checksum": "sha256:9f30b434e6c08d9a155bbc24ef6d71745d8d252c96d0ded0ccb6d158354c6546", "enabled": 1 } }, "lvm": { "100": { "checksum": "sha256:f0b2c006055d609858e59815fcd002fab6ce4a77761dfd2cc18b56219bf56fc1", "enabled": 1 } }, "mailman": { "100": { "checksum": "sha256:50901ef3e7ccc1828edc3fde37485af69afe28048ba25b7d9e3578b063d559d7", "enabled": 1 } }, "mailscanner": { "100": { "checksum": "sha256:914891424131853f76600695c57ec6283f9cc3c501c06e1427ef99f83431d1cd", "enabled": 1 } }, "man2html": { "100": { "checksum": "sha256:a7af8f4be03248a6d91c5df6c2becd22e3f45a71cecd671d44d2830fcd9f4a3b", "enabled": 1 } }, "mandb": { "100": { "checksum": "sha256:4afad481406029752a7960f97c253b41b8251018b219e16a858ccd029f9bb5f2", "enabled": 1 } }, "mcelog": { "100": { "checksum": "sha256:047b1501df33727c4257a655f1ad55aaed7da1d763782f7da45dfb1b19fa6a02", "enabled": 1 } }, "mediawiki": { "100": { "checksum": "sha256:7ce2990caacbf36b3ea6481def0630fa02e220ea9a4e2dc107251bce1368e569", "enabled": 1 } }, "memcached": { "100": { "checksum": "sha256:577518e3f4a348aaac570906781e3db46d9d53bf52d1a6f83c9f520dab8270e7", "enabled": 1 } }, "milter": { "100": { "checksum": "sha256:aa6812766cbdfad665abf5ad9633ffb945822d31714ffaa2bf4cb061f91e6b6c", "enabled": 1 } }, "minidlna": { "100": { "checksum": "sha256:f13c8b27b597e55d9e6342c756bc4d21de401fcbc4cb01e16c71c01a816c2e2b", "enabled": 1 } }, "minissdpd": { "100": { "checksum": "sha256:18649df6879516ea838eabbd19d6f7aefe2bb7da501c0b471c453d80ea4e02cb", "enabled": 1 } }, "miscfiles": { "100": { "checksum": "sha256:6f5792cde030e56b55f4b00cb53bd6d3724181c80aaaee5806e88c6a5298bb6c", "enabled": 1 } }, "mock": { "100": { "checksum": "sha256:3fdad88af11ebff86688d4db404640a1371af801e4a9b2f14a1b2abc1c76a503", "enabled": 1 } }, "modemmanager": { "100": { "checksum": "sha256:55370db1ae9a5e7ec0b4ec118f45ed61578656c6df3deb7b9e618362ddce4e29", "enabled": 1 } }, "modutils": { "100": { "checksum": "sha256:37b0e80e0d823ecd2606ece25c0b93eaea9f16da35261d79630b3e95c33a5fb4", "enabled": 1 } }, "mojomojo": { "100": { "checksum": "sha256:8ddaca2a54241adf5d42f9205cd073c7062e7692f31ebe15eb9e988e12efc9d4", "enabled": 1 } }, "mon_statd": { "100": { "checksum": "sha256:7e1bf20f24f949d86cf2cf1ecd59aef5aee94c9258d32a8fb4ee7646e4dc67e7", "enabled": 1 } }, "mongodb": { "100": { "checksum": "sha256:051c9eb43df88894edffecf5083ac5065a7051b743068b6f1866d923bca5ab95", "enabled": 1 } }, "motion": { "100": { "checksum": "sha256:154f6949eddb0f008a3b9166f736dde00bbc37df415f07c1263cd63ee609a5fd", "enabled": 1 } }, "mount": { "100": { "checksum": "sha256:acb9c37d19d0aec956cffe6f83d9593cb70bf8ad71ecf956e074b590286f489b", "enabled": 1 } }, "mozilla": { "100": { "checksum": "sha256:0a4982ad6ec48750689dc3c5d7ee946fca86c0c8b3763ee322aae33546109bed", "enabled": 1 } }, "mpd": { "100": { "checksum": "sha256:8807a444e6651ef4e67d0fed3bbc2558dc34b15e7293bfa03487d0a2c54ba905", "enabled": 1 } }, "mplayer": { "100": { "checksum": "sha256:3f999e549842f5a13f5f0d5aa4a752e74216c123f34f1797588b73d504afdc81", "enabled": 1 } }, "mptcpd": { "100": { "checksum": "sha256:f715709d773d8da3d99b4a7de877c7d7a8964d7a1342a61779edcdde2c75702d", "enabled": 1 } }, "mrtg": { "100": { "checksum": "sha256:740151127f1bcdb7699a6317eca571399e7cf58a755afd20dd141aec3035ab53", "enabled": 1 } }, "mta": { "100": { "checksum": "sha256:e04f6d7a528f9eeb7835428fe62d15043b9f4642cded56d682551e18920d04b3", "enabled": 1 } }, "munin": { "100": { "checksum": "sha256:1d2ea7b3403a2c521005ef980913298f01b2c00338366997273344590d0b218e", "enabled": 1 } }, "mysql": { "100": { "checksum": "sha256:07bc412265c70dcb7a3f85ba5265573ef76373364bbbd165ad3abb662e4de705", "enabled": 1 } }, "mythtv": { "100": { "checksum": "sha256:4cab8df79dc5a8e73ece06c418cb3ca58c74f669516b810de46b9f268d78ee8c", "enabled": 1 } }, "nagios": { "100": { "checksum": "sha256:712912fdf95fbf2adc1cb7427524d5536f0cd591bf3e77b63ef069213ae86704", "enabled": 1 } }, "namespace": { "100": { "checksum": "sha256:c032c4722dc3ec48bc9f196ae3de9ef9fd1c04554aa0913d99ade36ac0cf85af", "enabled": 1 } }, "ncftool": { "100": { "checksum": "sha256:e23fc9d15551fc1d7ba386350226d425282fcc090b9dfdcdb7065ebf3385666b", "enabled": 1 } }, "netlabel": { "100": { "checksum": "sha256:d9b7f3b8cb9b0ca564b43893085106dab4a6f6e7bee3143d778fc28218910435", "enabled": 1 } }, "netutils": { "100": { "checksum": "sha256:e828ed9b31086b0b53a5cf66cb2a580aae3e3290156b0d32c07eec5b1bd9cfe5", "enabled": 1 } }, "networkmanager": { "100": { "checksum": "sha256:bc61c1cee0f527dce7154fefb7906e6262e81d84108c7999f8db9cba7e4eb44c", "enabled": 1 } }, "ninfod": { "100": { "checksum": "sha256:4ed636e37871cc6a707bd29f6d990f4ff13d98bac2d8bf37922983d6289b3411", "enabled": 1 } }, "nis": { "100": { "checksum": "sha256:bce01b1750e64ca5f88e7d00d7d4dc88d5ebd430561f9b3cc680484b631c6c8a", "enabled": 1 } }, "nova": { "100": { "checksum": "sha256:16ebf87ef7e152147a0aa89e34161622d387348d4376f2eb91d971ef8ab476a0", "enabled": 1 } }, "nscd": { "100": { "checksum": "sha256:c7e5598679b39499dd00ac032b9a85b6e7e367bc64a0e2827cd799bb7e920a72", "enabled": 1 } }, "nsd": { "100": { "checksum": "sha256:75e019e0b2eda6179a3b39defbd43329dff4627e2c6a06c868ebee8b584332ee", "enabled": 1 } }, "nslcd": { "100": { "checksum": "sha256:69c6eaa2ef51784392d46e962c0f9b9439616dc97ba3105ff54e275a7b1e9db4", "enabled": 1 } }, "ntop": { "100": { "checksum": "sha256:cd4113e904b08b55f2a611ab02edbba284a91097a468b0d6b0d98a5dff397220", "enabled": 1 } }, "ntp": { "100": { "checksum": "sha256:5f188e0c5dabce04e39700cab0a3052cde82676b66e13b054d55f6f78f8d4b30", "enabled": 1 } }, "numad": { "100": { "checksum": "sha256:ed1193b412ae24e79c9a91531c99520b68ded7151bbf0f8cf0353d0dab713dec", "enabled": 1 } }, "nut": { "100": { "checksum": "sha256:54c04020d40b310f4b85f747395b3183859d816a6f181e4a54f5bd6f23b660df", "enabled": 1 } }, "nvme_stas": { "100": { "checksum": "sha256:1098f37f7b13bb6fa26d45bd4c54fd8cb38eddd1b4d02efaf88241f7338022c2", "enabled": 1 } }, "nx": { "100": { "checksum": "sha256:a8b41cc8424e4d9d67310c486221a9fc2701a9508104bdeb076f2ece6c94957c", "enabled": 1 } }, "obex": { "100": { "checksum": "sha256:4810ca21adc14e1f8b8f2a5b91e628038786e71ff84a2b3d701cdcafc6baf16c", "enabled": 1 } }, "oddjob": { "100": { "checksum": "sha256:d09199963de83937d060d3199ad48bf9f6753507c2748ad11d86628469f1e5b5", "enabled": 1 } }, "opafm": { "100": { "checksum": "sha256:020fbfc8bc4b762f26e66c399e2ae1f202c409d7ae1f10ef84f94c9672a1e55f", "enabled": 1 } }, "openct": { "100": { "checksum": "sha256:fd4ba492c3bc1b35ee1d9e4cf2b1fa21ac80ed31c9c76cb4503996f7c6367898", "enabled": 1 } }, "opendnssec": { "100": { "checksum": "sha256:cd5781c1d95be89112ad12c0d1341cc43b60259d4eea9b11442bc19da8ee462c", "enabled": 1 } }, "openfortivpn": { "100": { "checksum": "sha256:9a125b8b0328b3f01aea6aa395ce5117bc3a683e2b45b12e69dc6a95116ce203", "enabled": 1 } }, "openhpid": { "100": { "checksum": "sha256:65ec5e445fa5dd1471b83bccc319e9403a1c5888423442382751b1bf91c5188d", "enabled": 1 } }, "openshift": { "100": { "checksum": "sha256:07bea1192b1c51dd443db5a17d844ba3d7383abc3a3e7e979dcf0143753af074", "enabled": 1 } }, "openshift-origin": { "100": { "checksum": "sha256:88873289a3c1bd0c0c5d636b689bd8e1fe194ca528720c3191fb61b2ca024b50", "enabled": 1 } }, "opensm": { "100": { "checksum": "sha256:76a9f4d16f6b8fe2cb94b30b0c557dd4b067a4cb49502ff812b95290539062f2", "enabled": 1 } }, "openvpn": { "100": { "checksum": "sha256:ec0e41fdd7c04028cedf656a3fd60533adc778663a70734f5f2d569f5e79f53b", "enabled": 1 } }, "openvswitch": { "100": { "checksum": "sha256:1ab8ca3123bd0bbfbe2baac2e44124259401557c590cb8e906b06f28a4b9e32d", "enabled": 1 } }, "openwsman": { "100": { "checksum": "sha256:417b01a65501649535510d72e5fca7174183b5c873c1ddcb6921c052c8be3fc3", "enabled": 1 } }, "oracleasm": { "100": { "checksum": "sha256:6070bb82ad7b693e1ce70bcec6b975c4f5baec8a51e879f27e2c80ff2e076755", "enabled": 1 } }, "osad": { "100": { "checksum": "sha256:7bd10b072510204cc5f32e694b6eb1389fc20c40186fad86b76079e774023f4b", "enabled": 1 } }, "pads": { "100": { "checksum": "sha256:aeae1d20fad2ecc5e2f3ac84829becd50d70a7c41e8183c5b817e0d3f97f9484", "enabled": 1 } }, "passenger": { "100": { "checksum": "sha256:93cb4d1c0096f51325fa10c42ab72db38b5974c42b7d9fa6789482f56ce688c6", "enabled": 1 } }, "pcm": { "100": { "checksum": "sha256:4bb8d2c73a705f3051ce26155740950675bb556b11e2a7f59bb390ebbf7f9a87", "enabled": 1 } }, "pcmcia": { "100": { "checksum": "sha256:722b33d72298e2c8fa2900f1592c92673d098fcb6cd5802449086c918ca104dd", "enabled": 1 } }, "pcp": { "100": { "checksum": "sha256:a34e540529562d554136e3710fe67e8ca6e9ccee02b76dab369a448b58236f34", "enabled": 1 } }, "pcscd": { "100": { "checksum": "sha256:f50088e07d5fd52c3ad752e30852f0d3085cae9f13d2da6fd8c8141f1568ca28", "enabled": 1 } }, "pdns": { "100": { "checksum": "sha256:2d20aff28dd9cac707097c581edf455d754629927a32cc07854eb9baedb23b0a", "enabled": 1 } }, "pegasus": { "100": { "checksum": "sha256:81bedf6564a2a1e2b7ef03a9e5f646184f770202ad403e031f430e494abf8e88", "enabled": 1 } }, "permissivedomains": { "100": { "checksum": "sha256:2453bad4ace526f3cf2c60b358e95a5476692ef25da107b10f52f3af27c056d2", "enabled": 1 } }, "pesign": { "100": { "checksum": "sha256:544140731f1c09b504a6b85ec6ffbdda47c549c9858d6c2264498bfc56fd6127", "enabled": 1 } }, "pingd": { "100": { "checksum": "sha256:a4f1dd19ad8877fb1b48c431790bc350af44d3e2d0d08d59497d2a4425ce2d80", "enabled": 1 } }, "pkcs": { "100": { "checksum": "sha256:c0fe90911ac32abd43e43d9e25dd5a15c85301a1ad0ee579fd3997b9c25ad201", "enabled": 1 } }, "pki": { "100": { "checksum": "sha256:16249004873cf9214f949f73481b5476c0ab03e349542512af9b57aeb637c047", "enabled": 1 } }, "plymouthd": { "100": { "checksum": "sha256:4e33d3ebd3e3cd1fc439ca7bcf3c11086799c7af6a1e12460eb4fc5bab23b0be", "enabled": 1 } }, "podsleuth": { "100": { "checksum": "sha256:7f525e710a117aeaec78a9ab1cbbdd54a150ffe26ed5a1bc29595824fd9e1931", "enabled": 1 } }, "policykit": { "100": { "checksum": "sha256:fe85c9803d13edb28fd5a47803918af2a1e70d66b57e920fefe8e6cfdf778171", "enabled": 1 } }, "polipo": { "100": { "checksum": "sha256:8ab8b8f9438c6ae6c96323b6cf5937528d8223c38629f9bed4a2df00f4b9b686", "enabled": 1 } }, "portmap": { "100": { "checksum": "sha256:59c39477c07ba01841c150a2232e21074575ad3f93782e73b35da8f7acb121af", "enabled": 1 } }, "portreserve": { "100": { "checksum": "sha256:00aa972aa3ea3cf49a0134ebc75df0d6b7eb87ca7b31c61b3107001111abda8a", "enabled": 1 } }, "postfix": { "100": { "checksum": "sha256:68f178fd86c386a5d4df236f983239aa04eb6482304df3e9eb56340b744f5d8c", "enabled": 1 } }, "postgresql": { "100": { "checksum": "sha256:bab7ecb38d7aa40fc1de9d3f1a06e0f83c473e22e2f2391cbdc50e27ef92f7dd", "enabled": 1 } }, "postgrey": { "100": { "checksum": "sha256:8e1915326802644782f3041e65e3b1dfb6f534f2a66f4cb00ca8391f35ef2048", "enabled": 1 } }, "ppp": { "100": { "checksum": "sha256:912c8757c9d3cc46657aa135743c58836db26f4a6abf17807017c86a8512d90f", "enabled": 1 } }, "prelink": { "100": { "checksum": "sha256:0b0ad935c694badc8da9368226c53640cfade72c6828d094924029340ad5521e", "enabled": 1 } }, "prelude": { "100": { "checksum": "sha256:92c5efa0190469791dd3a105bebd81b2891596683001c6cb0d7446fc0d5125a2", "enabled": 1 } }, "privoxy": { "100": { "checksum": "sha256:0c0a1216baa5adcd33f85d92f248893c11cb55b5b068db990b2f94d9d7914932", "enabled": 1 } }, "procmail": { "100": { "checksum": "sha256:ab3962f174f9a9a125e2aad039584a83294b11d3ee56c8b477217a8193afc87d", "enabled": 1 } }, "prosody": { "100": { "checksum": "sha256:3573621c7a830f0d511041287fe8bb073eaa5e200c5b8b2e72096bf469c35f1d", "enabled": 1 } }, "psad": { "100": { "checksum": "sha256:72c77f99e96935aa093669bb095f914484b67e54ae66b4ef8ca780d2758cc6c8", "enabled": 1 } }, "ptchown": { "100": { "checksum": "sha256:318c62c2f63baefa43a61f8ef58bc168596087cd892a63afee72b1cb29cac0c7", "enabled": 1 } }, "pulseaudio": { "100": { "checksum": "sha256:3a7087d4680d1a35893caebf703937f36c1b1799dabc446b02428afc6c163c6f", "enabled": 1 } }, "puppet": { "100": { "checksum": "sha256:366806c68d03db1138cc9a3a08e0fdc4883825c20d1bed9a8e64ce7662bfb032", "enabled": 1 } }, "pwauth": { "100": { "checksum": "sha256:d23c293e39c1a025d375b714dd9319d27545c6c62b58342ef530763b84a24b16", "enabled": 1 } }, "qatlib": { "100": { "checksum": "sha256:cef00aa0bebb89745286c7db05374eb06fe9637c480d295f80f31d21be53672f", "enabled": 1 } }, "qmail": { "100": { "checksum": "sha256:5617289b46ae349fafd315cf92082d64119c6a0b5606bf79aa17fc99e34985b6", "enabled": 1 } }, "qpid": { "100": { "checksum": "sha256:1ee83ce320b057b0569c4bc9f8abdab6233eaf6fcd9c1a78975040bb2d9ddc56", "enabled": 1 } }, "quantum": { "100": { "checksum": "sha256:704e2c7d5fca4cb84e320b9359c6168037ebc87908330c0053c143c8821d3863", "enabled": 1 } }, "quota": { "100": { "checksum": "sha256:58dca8ea36c665ab74eb79cd51d54a7224514660bdfc4b666c2b8031e2877da3", "enabled": 1 } }, "rabbitmq": { "100": { "checksum": "sha256:acc51303259d5e6049265839c5ed13703baa80ae3477ae292103b4665c83dfbb", "enabled": 1 } }, "radius": { "100": { "checksum": "sha256:b53cd6e9f7493004e2b5bdf30e368fbc90b4f971db3fe56e55f31cc839969237", "enabled": 1 } }, "radvd": { "100": { "checksum": "sha256:c909b5d3b8ba285d71358dc7c9fca270e7b1586266c597a60b167ff600d10f7b", "enabled": 1 } }, "raid": { "100": { "checksum": "sha256:071d5ed18ad1c7d334bfba17d392c80a588ad00b1052ab0e21f9c53b0db95e4f", "enabled": 1 } }, "rasdaemon": { "100": { "checksum": "sha256:6d5c3ce954d1cef655ba323f1fef77bda8813815c0d991d020e0c760a4089aa5", "enabled": 1 } }, "rdisc": { "100": { "checksum": "sha256:3b10e49a0fec5da0e8998d77835464b25ced6ee9bd04a24dd861601a9aa02d2d", "enabled": 1 } }, "readahead": { "100": { "checksum": "sha256:e5b3d0b401606c60c1acdfce14c7b6b3f3363a1f58d8d91dd3be36e3df38005f", "enabled": 1 } }, "realmd": { "100": { "checksum": "sha256:b37f4de6bcd25f24eee675c18e2f1842023f0cf9a9a9a7923b64f7eaa875494f", "enabled": 1 } }, "redis": { "100": { "checksum": "sha256:b862f723b5285ad797d5a7546a17b1080690d9218a23b84bb0f7fe618169610f", "enabled": 1 } }, "remotelogin": { "100": { "checksum": "sha256:622d5ae5349b504c7ec9e472a86c5a1e1bc27606868d53d154112a63299c8dc9", "enabled": 1 } }, "restraint": { "400": { "checksum": "sha256:95d0d03fbc1d4147f02a0b3da7cd76efbdd75d1f5812cf6160e922336abbf270", "enabled": 1 } }, "rhcd": { "100": { "checksum": "sha256:fcd8c8415a58d937e726318252127c28e69af2819be01b977b4dda03994834a0", "enabled": 1 } }, "rhcs": { "100": { "checksum": "sha256:1e8298e1e453be52e1a86e391b827f2c71729c066f7d6c63abe7843e28647207", "enabled": 1 } }, "rhev": { "100": { "checksum": "sha256:eaa7e5657902e356ab7700656ff4c185eaf8df675d0a9a3f9ea1a4d8d8fb577d", "enabled": 1 } }, "rhgb": { "100": { "checksum": "sha256:5271469985f35fb7911a72e9e705bc35971294386838f1f11033657874821cea", "enabled": 1 } }, "rhnsd": { "100": { "checksum": "sha256:2a502ce16b7c56df48d0798ed7ca60370d0993ee425a542bf0621b51d4caf2d4", "enabled": 1 } }, "rhsmcertd": { "100": { "checksum": "sha256:80b01c4c3ddf665dd2fd5391b6069e9b868573f7c7a9d4421bc5bfeaadc71a22", "enabled": 1 } }, "rhts": { "400": { "checksum": "sha256:39f15722a115b11064c20b8bc2758e8fe06531a8f923ea00a3e7079a0554e3d6", "enabled": 1 } }, "ricci": { "100": { "checksum": "sha256:dd355c8a0a048baee894e0951365b525bbf5d7cce5907c9919bb0fffb1876661", "enabled": 1 } }, "rkhunter": { "100": { "checksum": "sha256:95ecb10dfe02e0384b0f220c46a7d0c85166a4e2cca0bb55654420b73611a646", "enabled": 1 } }, "rlogin": { "100": { "checksum": "sha256:d8594a73f246468d42569485bb148cbf1b582779074735f6da46e92f763984e1", "enabled": 1 } }, "rngd": { "100": { "checksum": "sha256:50b57a4776d2c6de8b1698bd31ecbfcc98d90d4a9bd2c95672b114c9637ce940", "enabled": 1 } }, "roundup": { "100": { "checksum": "sha256:54c5e85363d326f78c5896bf87c36fca9bc28acee3bd9da4d6d22f090dc9f7c6", "enabled": 1 } }, "rpc": { "100": { "checksum": "sha256:404f69ae03180ddce44c15efc636e5d7ff34b2108cb2c09624075be42a0545b1", "enabled": 1 } }, "rpcbind": { "100": { "checksum": "sha256:f5c703b4107b2616b1adf105b5c592c79f03df1f3a24f31453a242cdec50ea97", "enabled": 1 } }, "rpm": { "100": { "checksum": "sha256:bb52185531861597a0b44bca75497c08ac4c07e10ec63505c333bde800176fb7", "enabled": 1 } }, "rrdcached": { "100": { "checksum": "sha256:2cb11286a0afa94eccbd33de5a32b16d9e178e41ac1fec9dcd801b77dd257051", "enabled": 1 } }, "rshd": { "100": { "checksum": "sha256:0052600f999e54f224d1210c5920b42227cd98ce8db24155b0cfecdf4f6c6474", "enabled": 1 } }, "rshim": { "100": { "checksum": "sha256:7825528bbcdb081ead1d0d2da0d5156e430f93d003f58dcebbb8997c906f0dec", "enabled": 1 } }, "rssh": { "100": { "checksum": "sha256:cde9bfc64b345835e9f8012426cf218bad9972c70e22e0ef6d6a90cc5fa7163c", "enabled": 1 } }, "rsync": { "100": { "checksum": "sha256:1295d1a5434aa1df7fd0f0e2a9b2b045b0ec51bd8619d528349530b322eddf24", "enabled": 1 } }, "rtas": { "100": { "checksum": "sha256:8c5c3a495b42ed14f56e360dd9eaaa0df2426f2c93676880f129bab95e247338", "enabled": 1 } }, "rtkit": { "100": { "checksum": "sha256:7b0c8b0fd23723aed98bc581f2ead99da87893c6ba042ec19ca731e1d802eab6", "enabled": 1 } }, "rwho": { "100": { "checksum": "sha256:fe02b4fb75b8a75ac77b53ae13939199422543f68e71829fd1530aa34c8715cc", "enabled": 1 } }, "samba": { "100": { "checksum": "sha256:b0af5e99697f8e864538813653ea879ea50b6a9269424ca1deaa95ec1a4a7778", "enabled": 1 } }, "sambagui": { "100": { "checksum": "sha256:cde09885f6d657107c517f09f01933816199f0694799fe3df4ac281b411edbb9", "enabled": 1 } }, "sandboxX": { "100": { "checksum": "sha256:ffd79700483dc635b0396df2f9ca75062aff985e7874c2014bca95e80b08bd5d", "enabled": 1 } }, "sanlock": { "100": { "checksum": "sha256:fcd9cea4d61e3597a5e5cb1bd36efc228a21e5b8ee79c68b44e8a93aae834fb6", "enabled": 1 } }, "sap": { "100": { "checksum": "sha256:4cffd87a27c2b48cc59f596a42e860ea52acdfcbff0c2a1e4aeafa3338b29e6e", "enabled": 1 } }, "sasl": { "100": { "checksum": "sha256:ed40cd297b033fd86770692e7d3ab76598684f521f079475d8b44953116291cf", "enabled": 1 } }, "sbd": { "100": { "checksum": "sha256:8c40daa2d8288e7535e9db7724d16a34add080c65fab32bdf5e10cdace95a063", "enabled": 1 } }, "sblim": { "100": { "checksum": "sha256:73b934662023f728bfeb5cac500547e15e389bad3575a3750bf05901d4396531", "enabled": 1 } }, "screen": { "100": { "checksum": "sha256:8dcba8c325fdb2c56f7c7c77362e287469484b19f0b491dd7bc0abc6e033378b", "enabled": 1 } }, "secadm": { "100": { "checksum": "sha256:2d9e07e5ee506dbf6cf976e257210fdb14222c1e437b1f7dceb1548fa399beba", "enabled": 1 } }, "sectoolm": { "100": { "checksum": "sha256:e6c4cabc853877f459381f06d4d47660ce9360a188ab87bb825f1f1eaed39d7d", "enabled": 1 } }, "selinuxutil": { "100": { "checksum": "sha256:38da361f759758e151b7f8612954c91c0fde09615e1c82242c80706231e16355", "enabled": 1 } }, "sendmail": { "100": { "checksum": "sha256:286d18507d80fcfa7cb83ed21321df23c3a29001777890137c1b52c113cd8489", "enabled": 1 } }, "sensord": { "100": { "checksum": "sha256:6681fdfffe9ba81fa3712f665b98feb68f9331f8f282faf19a4ae5e70b02bd63", "enabled": 1 } }, "setrans": { "100": { "checksum": "sha256:69a3bce8a9e35afa4c3b4c7e2f4ba604491313a688d36e3158f96e90343ea370", "enabled": 1 } }, "setroubleshoot": { "100": { "checksum": "sha256:c0d9c0c6cea100697cbc684f2ab74ec66d7c121ca88200752585aeaf4c2b2596", "enabled": 1 } }, "seunshare": { "100": { "checksum": "sha256:219912d16220957b98bbe694f06c3509d2f884ff8418da1eeeda43f4dce81dc9", "enabled": 1 } }, "shorewall": { "100": { "checksum": "sha256:d0dd25c1a5b77a28fcf61f4796c479a026a8e2ec944f5e4d1492cd552aa9ca02", "enabled": 1 } }, "slocate": { "100": { "checksum": "sha256:8d149bca3982ac8c0f7376571d1f0614f05eb2fe7465951eefb17153f642ee40", "enabled": 1 } }, "slpd": { "100": { "checksum": "sha256:6798797c2dda6b8b24aada5a4bbc4f8775faca22435ed1de39193457043f8d8b", "enabled": 1 } }, "smartmon": { "100": { "checksum": "sha256:9de88cb8f60049e56dc5ead706a324189ff276929ccc3bf4ba7d06e96c0d8414", "enabled": 1 } }, "smokeping": { "100": { "checksum": "sha256:29560964ba6e3cb5fc392417ff4aa3e50a5f15c2bc21bdf37088c87868ffeec7", "enabled": 1 } }, "smoltclient": { "100": { "checksum": "sha256:0e47cf6e775a1e2ecb506e444731e086d057b7f2c2aeb6bce7b636372de88d95", "enabled": 1 } }, "snapper": { "100": { "checksum": "sha256:e264d81b3e1173c98c1efc90529d7f9c2c9047b843e29a4fb807d0b1b96b4685", "enabled": 1 } }, "snmp": { "100": { "checksum": "sha256:ec6638b69ec04510a807a5d2c97568a670f80a7999ee711e5e1874136923718e", "enabled": 1 } }, "snort": { "100": { "checksum": "sha256:67c1751c124c249fa15c36f948b60a76db3729f819678cecba569c79abec3af2", "enabled": 1 } }, "sosreport": { "100": { "checksum": "sha256:345531e273b1a0c0d6e02a1898d60759007d5126ec2016eea320832c33ac8211", "enabled": 1 } }, "soundserver": { "100": { "checksum": "sha256:932d162b8365f213b97eb6a7e62d20b8c30bfe1863b6534462da96077d6347fc", "enabled": 1 } }, "spamassassin": { "100": { "checksum": "sha256:fbceb12b459776c22867efa6bf2d07657e058c4a80d6f20e724086ec58e600b0", "enabled": 1 } }, "speech-dispatcher": { "100": { "checksum": "sha256:092593b379ff3e7b3d15b0dfdd2adfde73fb76401d1b6b6e79a4efccbf03f0c9", "enabled": 1 } }, "squid": { "100": { "checksum": "sha256:4374a853cb79bd24d05f65585ad1a74a7b9da4d3454ca233da351e009e75ed79", "enabled": 1 } }, "ssh": { "100": { "checksum": "sha256:493ee98f9cda10ca995eb5600d2940556d70cb3f544501c6bd1968745cc234fa", "enabled": 1 } }, "sslh": { "100": { "checksum": "sha256:5dc03f6d33a6caa7c2fb5ede0f8ded4ecb8eb244443614189cb5909521cab539", "enabled": 1 } }, "sssd": { "100": { "checksum": "sha256:4ee157395107db3b146239f14c9868f6e9d71199520e5d2e473e5e836d0635d1", "enabled": 1 } }, "staff": { "100": { "checksum": "sha256:6a89d52653ce210b1dfb2d2b10e184cacb18976d238fc92b349c40661fdcbf47", "enabled": 1 } }, "stalld": { "100": { "checksum": "sha256:14aa28921885421f92fb51920ca0d8fde122881ac9e28d024c80a7b234b82dd3", "enabled": 1 } }, "stapserver": { "100": { "checksum": "sha256:1610ffc61426c0e791bd7a6a2628935edf3d66fc7818fb42c314044a26ac5d2d", "enabled": 1 } }, "stratisd": { "100": { "checksum": "sha256:5025b28ad7090d4802078c92ba257b46909bbb74fcd6cae0f2a5d65978a5ff8e", "enabled": 1 } }, "stunnel": { "100": { "checksum": "sha256:67b1c960e0cd11cdf1446615cf633700585e69f41c4039cfb183d69527648f1e", "enabled": 1 } }, "su": { "100": { "checksum": "sha256:60850779bd04d036ea792653d4e794f0e865936648c03030515d8125cf412b4b", "enabled": 1 } }, "sudo": { "100": { "checksum": "sha256:0965d3f30b2813a4364e9ce8da0f8916579319ed9507640383c3bf195ae1c00e", "enabled": 1 } }, "svnserve": { "100": { "checksum": "sha256:d589d2e8bddbcea03272bb589cffde776ccb410ac685e5071896fa8e59660c7f", "enabled": 1 } }, "swift": { "100": { "checksum": "sha256:407ddd533381c1f822ec8471e639404c4b926e7e635b3d25d4ba1403a5c517b3", "enabled": 1 } }, "sysadm": { "100": { "checksum": "sha256:1d310d8df59d431579fa3568b7ac14f853223aadd4087fe4c57a1add6606987d", "enabled": 1 } }, "sysadm_secadm": { "100": { "checksum": "sha256:6a0a3b3c4028f304beeacb7d06c62975a2bfb13d67a6a4fe6b95f6d6eeaa74b7", "enabled": 1 } }, "sysnetwork": { "100": { "checksum": "sha256:56b999cfaf7defa2c5cdd5c0c16e7c9c1de940dcc528778ff752fca51f95b38a", "enabled": 1 } }, "sysstat": { "100": { "checksum": "sha256:358b9c77d350c4bc17cf773e99f81d5267f8c41d47402234be37dd9f31655ad0", "enabled": 1 } }, "systemd": { "100": { "checksum": "sha256:124460467ec0f3f7443a67583ef4613154b2b92615e0d779a3ec8060c5a8ce4a", "enabled": 1 } }, "tangd": { "100": { "checksum": "sha256:a2e2a5d807eabe30a09085dcec38d530781c1f02e840a6ed02ccc32bd2738e60", "enabled": 1 } }, "targetd": { "100": { "checksum": "sha256:4a9e67698b4759a406188562c375ecd847f205c38873fb895c3c20abc8bac869", "enabled": 1 } }, "tcpd": { "100": { "checksum": "sha256:b69263d84bf511825be889b667401c30964c7d30d1147d577ae294cd7f05aa01", "enabled": 1 } }, "tcsd": { "100": { "checksum": "sha256:fa40e623b09fb5f61c2c2113a8ed0d17a4a5d105b9f22389ce088a766c0042c0", "enabled": 1 } }, "telepathy": { "100": { "checksum": "sha256:d98021b3bbdd6255c2009e5adafbf41faf3370434396cf2e928e7df693dd50b6", "enabled": 1 } }, "telnet": { "100": { "checksum": "sha256:abf90c09c67553f2b04cab7461138bfd8689752a752a79e36e248f209eae73e6", "enabled": 1 } }, "tftp": { "100": { "checksum": "sha256:f5dcde99c32caa66ea23a9be3b2d6d95550eff7db62b11733f16f1f5eb01bce2", "enabled": 1 } }, "tgtd": { "100": { "checksum": "sha256:6a96f6d7abaa0e1951832347e924c5aff5c66194da3881ec4855e1401a5ce6d2", "enabled": 1 } }, "thin": { "100": { "checksum": "sha256:b0c1055c6a8419035ee9a9a45de2bcdd7183a906bb40d98bb2be80b882a60c51", "enabled": 1 } }, "thumb": { "100": { "checksum": "sha256:928596446716a5d1ef1f8958d150959e59a6258975ebf453c2d5a2747e6d59fd", "enabled": 1 } }, "tlp": { "100": { "checksum": "sha256:540f0a0339feec1e11d145c6ba99f700750451a2fad43190e256e590c3835950", "enabled": 1 } }, "tmpreaper": { "100": { "checksum": "sha256:4370bc58da3c779088151878e7291e03fcd2ed706425b4f2eeeed7097f8adcad", "enabled": 1 } }, "tomcat": { "100": { "checksum": "sha256:98e40c398d94d82eea102055d52f59f77581955b0e463dd95386b4559a4cc852", "enabled": 1 } }, "tor": { "100": { "checksum": "sha256:b309ee2c391d91ace45662d2152870961635ca8968487d0f2570508e160dbd9b", "enabled": 1 } }, "tuned": { "100": { "checksum": "sha256:90bcbb5cad47c032a02390e0c4b22ff79f93fbfd2eb4a385441a87532d6aa632", "enabled": 1 } }, "tvtime": { "100": { "checksum": "sha256:a4438b3063cc1760c663e35a7be70a7dcd099de505ce1562e02090ac6b635787", "enabled": 1 } }, "udev": { "100": { "checksum": "sha256:f308ba2a718171dc871e2586f35a90cb1bbc37decf13e714af72b38cc44505a2", "enabled": 1 } }, "ulogd": { "100": { "checksum": "sha256:05aa3f423e0505dfef9b5e311b0f0a71d93159c7f6440979d3c00a5d9d22ef4c", "enabled": 1 } }, "uml": { "100": { "checksum": "sha256:e0257dd4e250c5aa0baa8151d48942735949bd3018318c11c28a4a5df2aacb90", "enabled": 1 } }, "unconfined": { "100": { "checksum": "sha256:26c4304a314b04f6269b45b4285b449caad98d22de7b3017ef648212c6674da1", "enabled": 1 } }, "unconfineduser": { "100": { "checksum": "sha256:af2778eed6f3cc25c5390e4b1a892995c37a83f5952f51820007b8cd387833b4", "enabled": 1 } }, "unlabelednet": { "100": { "checksum": "sha256:c1de2adfab2cd577a0cc9f2c1a50cd6321324d31b9c3f81be9231f85d4acbbd3", "enabled": 1 } }, "unprivuser": { "100": { "checksum": "sha256:dc1ddbf10a94d55c476ff343c862c24dc649d82190a16a0fd924c527743a667d", "enabled": 1 } }, "updfstab": { "100": { "checksum": "sha256:11eb931e80afefafd0f1c41ea19add0c746170447c76dd629146c43bd9e98067", "enabled": 1 } }, "usbmodules": { "100": { "checksum": "sha256:3a8334855e0cfe67e2c29198baacf61ea0e397d5da30f60fb844044cd105557e", "enabled": 1 } }, "usbmuxd": { "100": { "checksum": "sha256:e883ad3abf2e2792abf4122f99db904139b2f1833b7d99c45b34a7c8244d7c66", "enabled": 1 } }, "userdomain": { "100": { "checksum": "sha256:b6cc74d18c0d17741092aaa8e830f160021ad38bdc38880e3b1488ef30f59a49", "enabled": 1 } }, "userhelper": { "100": { "checksum": "sha256:0a80be6b1635de7593f46421219be5a6ccb9c22973c60b74436ba16884d789fc", "enabled": 1 } }, "usermanage": { "100": { "checksum": "sha256:c56dfa32f163d476dd5761e762ea743189a97636acaf2b4eefcf88812e624877", "enabled": 1 } }, "usernetctl": { "100": { "checksum": "sha256:9c0430aca661392960008075f0a1a934f0c05239976e73c4f0408e9cb5e279ca", "enabled": 1 } }, "uucp": { "100": { "checksum": "sha256:527db920542a250b24e24cb8c6bd86502c3be7e270d11f43d4ecb1a6632f8421", "enabled": 1 } }, "uuidd": { "100": { "checksum": "sha256:b3df65ab2d82cbb05e872f00f46156cfbae937a3a6f850d05f43f4866b772af8", "enabled": 1 } }, "varnishd": { "100": { "checksum": "sha256:9174c3c68bf6b4c2b2b9be7e2a0d985ce6edfdec0dc4c225b427b986449d3e7a", "enabled": 1 } }, "vdagent": { "100": { "checksum": "sha256:52636260af5ae220698d505c6a90a33a71f5878a2cea3a303033c896f000b732", "enabled": 1 } }, "vhostmd": { "100": { "checksum": "sha256:df2d64343b871c840603e82f5fc1f7928c0345b314450e536f9ef66becf131ad", "enabled": 1 } }, "virt": { "100": { "checksum": "sha256:c209f246ed99c66defe424c0f1bd13e970df31d84d2f4e4a2a79ef1c00c3bbe2", "enabled": 1 } }, "virt_supplementary": { "100": { "checksum": "sha256:e8147b1556404b4a2ce521cc4c22e28d4ee42045ea2480043fc8bd2d5a593207", "enabled": 1 } }, "vlock": { "100": { "checksum": "sha256:54a1b28cebc1928622edc4f1f1762a5f2d57783be153052d592f2d10dd9b24a1", "enabled": 1 } }, "vmtools": { "100": { "checksum": "sha256:323b80c9ba1efb9852725d635a5aa442ae2bbd4ec74360cf75853616401dd4f9", "enabled": 1 } }, "vmware": { "100": { "checksum": "sha256:b68b94c1c4662f7005faac011c887619dbb76d69d3d2b63f354d48a144d0d63c", "enabled": 1 } }, "vnstatd": { "100": { "checksum": "sha256:c544da94ee903d411f7ab352b3fe286fa0a1b0fb8490d508565df66d5501b852", "enabled": 1 } }, "vpn": { "100": { "checksum": "sha256:035f214c319dcee2ce4e9751fa9cf1343d81d569409f3a56faf79e506e445504", "enabled": 1 } }, "w3c": { "100": { "checksum": "sha256:d6c6c226dcc92d9a668f42b86dde398d2d875b479915af656734357e20ed0e96", "enabled": 1 } }, "watchdog": { "100": { "checksum": "sha256:a237aeaf94bed19de41ef5289be96ce5bf85dce6f2fc4f9d36d6185b4e59520a", "enabled": 1 } }, "wdmd": { "100": { "checksum": "sha256:1012fdefa5dcf2e540f536de5da79df0ba715ccc2a86a414e2e846675d32495f", "enabled": 1 } }, "webadm": { "100": { "checksum": "sha256:f374fcd23ed5f30ce1159470af1e1254e0352a6ad7cf261287de8ec51f2b2626", "enabled": 1 } }, "webalizer": { "100": { "checksum": "sha256:b66dfdfd23ed304d5f416bb0d8de4565d28cd732253fc2f9e4b4bac59515790f", "enabled": 1 } }, "wine": { "100": { "checksum": "sha256:a4a3be2a0feea4de24ca26c0eed58c7025304d810ee9d4a3e1b7e964a8d28c6d", "enabled": 1 } }, "wireguard": { "100": { "checksum": "sha256:3084b8fd20df45b718cb6779240f4292a0437f536bde7db3616c40b1500ba0f6", "enabled": 1 } }, "wireshark": { "100": { "checksum": "sha256:3ec8e55883a8b73de513f533ed80b1b665e3ed650fdf2b471e5565e1cc3034d9", "enabled": 1 } }, "xen": { "100": { "checksum": "sha256:adabf99dccce9182074a23fb8f05ac537899170fbbfe022f39cb520275339186", "enabled": 1 } }, "xguest": { "100": { "checksum": "sha256:8a7f7d9514fa5257088e451dcf96c6c221a7b6499cae9ac15a688cbe2e84def9", "enabled": 1 } }, "xserver": { "100": { "checksum": "sha256:ca8b72b391320d267580cebf2ad9ef811a781b34647e1c63e525e982911cf956", "enabled": 1 } }, "zabbix": { "100": { "checksum": "sha256:ed9800f41440cf137baa7864acdd8a503b7e051cf023d252b87daeb5357be889", "enabled": 1 } }, "zarafa": { "100": { "checksum": "sha256:c3a9d1c95f377fafb9e94a84bec89645a1053d23b557b143367992eae9993596", "enabled": 1 } }, "zebra": { "100": { "checksum": "sha256:77d57a56a3196f09ea5eb808f8c3b4cef333b241e9007b91779ba78a857e5cfa", "enabled": 1 } }, "zoneminder": { "100": { "checksum": "sha256:675bdf525f863eaf17fe6c25507cf2dc3c4ab05d1e4b0d3477573d34980a5cd2", "enabled": 1 } }, "zosremote": { "100": { "checksum": "sha256:79bb5490f9f16e8caff67fc87a834bf6f11647cfed990c421616eb1adaac5c70", "enabled": 1 } } }, "selinux_priorities": true }, "changed": false } TASK [fedora.linux_system_roles.selinux : Load SELinux modules] **************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:115 Saturday 11 January 2025 11:30:58 -0500 (0:00:03.072) 0:02:23.271 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "selinux_modules is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.selinux : Restore SELinux labels on filesystem tree] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:128 Saturday 11 January 2025 11:30:58 -0500 (0:00:00.160) 0:02:23.432 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.selinux : Restore SELinux labels on filesystem tree in check mode] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:136 Saturday 11 January 2025 11:30:58 -0500 (0:00:00.060) 0:02:23.492 ****** skipping: [managed-node2] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:153 Saturday 11 January 2025 11:30:58 -0500 (0:00:00.104) 0:02:23.597 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:157 Saturday 11 January 2025 11:30:58 -0500 (0:00:00.063) 0:02:23.661 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:166 Saturday 11 January 2025 11:30:58 -0500 (0:00:00.058) 0:02:23.719 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:175 Saturday 11 January 2025 11:30:58 -0500 (0:00:00.063) 0:02:23.783 ****** skipping: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:182 Saturday 11 January 2025 11:30:58 -0500 (0:00:00.060) 0:02:23.843 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml for managed-node2 => (item=(censored due to no_log)) included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml for managed-node2 => (item=(censored due to no_log)) included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml for managed-node2 => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:14 Saturday 11 January 2025 11:30:59 -0500 (0:00:00.271) 0:02:24.115 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_spec": { "debug": true, "log_level": "debug", "state": "started" }, "__podman_kube_str": "apiVersion: v1\nkind: Pod\nmetadata:\n labels:\n app: test\n io.containers.autoupdate: registry\n name: httpd1\nspec:\n containers:\n - command:\n - /bin/busybox-extras\n - httpd\n - -f\n - -p\n - 80\n image: quay.io/libpod/testimage:20210610\n name: httpd1\n ports:\n - containerPort: 80\n hostPort: 15001\n volumeMounts:\n - mountPath: /var/www:Z\n name: www\n - mountPath: /var/httpd-create:Z\n name: create\n workingDir: /var/www\n volumes:\n - hostPath:\n path: /tmp/lsr_adhv0_3l_podman/httpd1\n name: www\n - hostPath:\n path: /tmp/lsr_adhv0_3l_podman/httpd1-create\n name: create\n" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:21 Saturday 11 January 2025 11:30:59 -0500 (0:00:00.097) 0:02:24.213 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_kube": { "apiVersion": "v1", "kind": "Pod", "metadata": { "labels": { "app": "test", "io.containers.autoupdate": "registry" }, "name": "httpd1" }, "spec": { "containers": [ { "command": [ "/bin/busybox-extras", "httpd", "-f", "-p", 80 ], "image": "quay.io/libpod/testimage:20210610", "name": "httpd1", "ports": [ { "containerPort": 80, "hostPort": 15001 } ], "volumeMounts": [ { "mountPath": "/var/www:Z", "name": "www" }, { "mountPath": "/var/httpd-create:Z", "name": "create" } ], "workingDir": "/var/www" } ], "volumes": [ { "hostPath": { "path": "/tmp/lsr_adhv0_3l_podman/httpd1" }, "name": "www" }, { "hostPath": { "path": "/tmp/lsr_adhv0_3l_podman/httpd1-create" }, "name": "create" } ] } }, "__podman_kube_file": "", "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "podman_basic_user" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:33 Saturday 11 January 2025 11:30:59 -0500 (0:00:00.167) 0:02:24.380 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_name": "httpd1", "__podman_rootless": true }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:38 Saturday 11 January 2025 11:30:59 -0500 (0:00:00.069) 0:02:24.449 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 11 January 2025 11:30:59 -0500 (0:00:00.124) 0:02:24.573 ****** ok: [managed-node2] => { "ansible_facts": { "getent_passwd": { "podman_basic_user": [ "x", "3001", "3001", "", "/home/podman_basic_user", "/bin/bash" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 11 January 2025 11:31:00 -0500 (0:00:00.423) 0:02:24.997 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 11 January 2025 11:31:00 -0500 (0:00:00.067) 0:02:25.065 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_group": "3001" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Saturday 11 January 2025 11:31:00 -0500 (0:00:00.058) 0:02:25.123 ****** ok: [managed-node2] => { "changed": false, "stat": { "atime": 1736612945.6352425, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1736612940.9742577, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 9160785, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "1643853349", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Saturday 11 January 2025 11:31:00 -0500 (0:00:00.412) 0:02:25.535 ****** ok: [managed-node2] => { "changed": false, "cmd": [ "getsubids", "podman_basic_user" ], "delta": "0:00:00.004540", "end": "2025-01-11 11:31:00.895460", "rc": 0, "start": "2025-01-11 11:31:00.890920" } STDOUT: 0: podman_basic_user 524288 65536 TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Saturday 11 January 2025 11:31:00 -0500 (0:00:00.421) 0:02:25.957 ****** ok: [managed-node2] => { "changed": false, "cmd": [ "getsubids", "-g", "podman_basic_user" ], "delta": "0:00:00.005010", "end": "2025-01-11 11:31:01.292854", "rc": 0, "start": "2025-01-11 11:31:01.287844" } STDOUT: 0: podman_basic_user 524288 65536 TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Saturday 11 January 2025 11:31:01 -0500 (0:00:00.422) 0:02:26.379 ****** ok: [managed-node2] => { "ansible_facts": { "podman_subgid_info": { "podman_basic_user": { "range": 65536, "start": 524288 } }, "podman_subuid_info": { "podman_basic_user": { "range": 65536, "start": 524288 } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Saturday 11 January 2025 11:31:01 -0500 (0:00:00.096) 0:02:26.475 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Saturday 11 January 2025 11:31:01 -0500 (0:00:00.069) 0:02:26.545 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Saturday 11 January 2025 11:31:01 -0500 (0:00:00.175) 0:02:26.720 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Saturday 11 January 2025 11:31:01 -0500 (0:00:00.064) 0:02:26.784 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Saturday 11 January 2025 11:31:01 -0500 (0:00:00.048) 0:02:26.832 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if no kube spec is given] ******** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:43 Saturday 11 January 2025 11:31:01 -0500 (0:00:00.052) 0:02:26.885 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "not __podman_kube", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:52 Saturday 11 January 2025 11:31:01 -0500 (0:00:00.043) 0:02:26.928 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_systemd_scope": "user", "__podman_user_home_dir": "/home/podman_basic_user", "__podman_xdg_runtime_dir": "/run/user/3001" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:60 Saturday 11 January 2025 11:31:02 -0500 (0:00:00.061) 0:02:26.989 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_path": "/home/podman_basic_user/.config/containers/ansible-kubernetes.d" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:64 Saturday 11 January 2025 11:31:02 -0500 (0:00:00.043) 0:02:27.033 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_kube_file": "/home/podman_basic_user/.config/containers/ansible-kubernetes.d/httpd1.yml" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:68 Saturday 11 January 2025 11:31:02 -0500 (0:00:00.049) 0:02:27.082 ****** ok: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Get service name using systemd-escape] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:75 Saturday 11 January 2025 11:31:02 -0500 (0:00:00.099) 0:02:27.182 ****** ok: [managed-node2] => { "changed": false, "cmd": [ "systemd-escape", "--template", "podman-kube@.service", "/home/podman_basic_user/.config/containers/ansible-kubernetes.d/httpd1.yml" ], "delta": "0:00:00.005290", "end": "2025-01-11 11:31:02.537984", "rc": 0, "start": "2025-01-11 11:31:02.532694" } STDOUT: podman-kube@-home-podman_basic_user-.config-containers-ansible\x2dkubernetes.d-httpd1.yml.service TASK [fedora.linux_system_roles.podman : Cleanup containers and services] ****** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:83 Saturday 11 January 2025 11:31:02 -0500 (0:00:00.446) 0:02:27.628 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update containers and services] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_kube_spec.yml:87 Saturday 11 January 2025 11:31:02 -0500 (0:00:00.067) 0:02:27.696 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:2 Saturday 11 January 2025 11:31:02 -0500 (0:00:00.259) 0:02:27.955 ****** included: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node2 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 11 January 2025 11:31:03 -0500 (0:00:00.121) 0:02:28.076 ****** changed: [managed-node2] => { "changed": true, "cmd": [ "loginctl", "enable-linger", "podman_basic_user" ], "delta": "0:00:00.019969", "end": "2025-01-11 11:31:03.462444", "rc": 0, "start": "2025-01-11 11:31:03.442475" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 11 January 2025 11:31:03 -0500 (0:00:00.504) 0:02:28.580 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 11 January 2025 11:31:03 -0500 (0:00:00.080) 0:02:28.661 ****** skipping: [managed-node2] => { "changed": false, "false_condition": "__podman_item_state | d('present') == 'absent'", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the host mount volumes] *********** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:7 Saturday 11 January 2025 11:31:03 -0500 (0:00:00.074) 0:02:28.735 ****** ok: [managed-node2] => { "ansible_facts": { "__podman_volumes": [ "/tmp/lsr_adhv0_3l_podman/httpd1", "/tmp/lsr_adhv0_3l_podman/httpd1-create" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:18 Saturday 11 January 2025 11:31:03 -0500 (0:00:00.096) 0:02:28.831 ****** changed: [managed-node2] => (item=/tmp/lsr_adhv0_3l_podman/httpd1) => { "ansible_loop_var": "item", "changed": true, "gid": 3001, "group": "podman_basic_user", "item": "/tmp/lsr_adhv0_3l_podman/httpd1", "mode": "0755", "owner": "podman_basic_user", "path": "/tmp/lsr_adhv0_3l_podman/httpd1", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 23, "state": "directory", "uid": 3001 } changed: [managed-node2] => (item=/tmp/lsr_adhv0_3l_podman/httpd1-create) => { "ansible_loop_var": "item", "changed": true, "gid": 3001, "group": "podman_basic_user", "item": "/tmp/lsr_adhv0_3l_podman/httpd1-create", "mode": "0755", "owner": "podman_basic_user", "path": "/tmp/lsr_adhv0_3l_podman/httpd1-create", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 3001 } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:29 Saturday 11 January 2025 11:31:04 -0500 (0:00:00.862) 0:02:29.694 ****** changed: [managed-node2] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [managed-node2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Check the kubernetes yaml file] ******* task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:53 Saturday 11 January 2025 11:31:06 -0500 (0:00:02.252) 0:02:31.947 ****** ok: [managed-node2] => { "changed": false, "failed_when_result": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Ensure the kubernetes directory is present] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:61 Saturday 11 January 2025 11:31:07 -0500 (0:00:00.406) 0:02:32.353 ****** changed: [managed-node2] => { "changed": true, "gid": 3001, "group": "podman_basic_user", "mode": "0755", "owner": "podman_basic_user", "path": "/home/podman_basic_user/.config/containers/ansible-kubernetes.d", "secontext": "unconfined_u:object_r:config_home_t:s0", "size": 6, "state": "directory", "uid": 3001 } TASK [fedora.linux_system_roles.podman : Ensure kubernetes yaml files are present] *** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:70 Saturday 11 January 2025 11:31:07 -0500 (0:00:00.417) 0:02:32.771 ****** changed: [managed-node2] => { "changed": true, "checksum": "bd406dc7744755fdf41f83e27a5ef8497bec46ba", "dest": "/home/podman_basic_user/.config/containers/ansible-kubernetes.d/httpd1.yml", "gid": 3001, "group": "podman_basic_user", "md5sum": "9f2e767e65ede611d1c6ccadb6e83852", "mode": "0644", "owner": "podman_basic_user", "secontext": "unconfined_u:object_r:config_home_t:s0", "size": 762, "src": "/root/.ansible/tmp/ansible-tmp-1736613067.8332987-12432-130581053166951/.source.yml", "state": "file", "uid": 3001 } TASK [fedora.linux_system_roles.podman : Update containers/pods] *************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:80 Saturday 11 January 2025 11:31:08 -0500 (0:00:00.712) 0:02:33.483 ****** fatal: [managed-node2]: FAILED! => { "changed": false } MSG: Output: [starting container 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de: cannot get namespace path unless container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e is running: container is stopped] Pod: cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da Container: 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de Error=time="2025-01-11T11:31:08-05:00" level=info msg="/bin/podman filtering at log level debug" time="2025-01-11T11:31:08-05:00" level=debug msg="Called kube.PersistentPreRunE(/bin/podman play kube --start=true --log-level=debug /home/podman_basic_user/.config/containers/ansible-kubernetes.d/httpd1.yml)" time="2025-01-11T11:31:08-05:00" level=debug msg="Using conmon: \"/usr/bin/conmon\"" time="2025-01-11T11:31:08-05:00" level=info msg="Using sqlite as database backend" time="2025-01-11T11:31:08-05:00" level=debug msg="systemd-logind: Unknown object '/'." time="2025-01-11T11:31:08-05:00" level=debug msg="Using graph driver overlay" time="2025-01-11T11:31:08-05:00" level=debug msg="Using graph root /home/podman_basic_user/.local/share/containers/storage" time="2025-01-11T11:31:08-05:00" level=debug msg="Using run root /run/user/3001/containers" time="2025-01-11T11:31:08-05:00" level=debug msg="Using static dir /home/podman_basic_user/.local/share/containers/storage/libpod" time="2025-01-11T11:31:08-05:00" level=debug msg="Using tmp dir /run/user/3001/libpod/tmp" time="2025-01-11T11:31:08-05:00" level=debug msg="Using volume path /home/podman_basic_user/.local/share/containers/storage/volumes" time="2025-01-11T11:31:08-05:00" level=debug msg="Using transient store: false" time="2025-01-11T11:31:08-05:00" level=debug msg="[graphdriver] trying provided driver \"overlay\"" time="2025-01-11T11:31:08-05:00" level=debug msg="Cached value indicated that overlay is supported" time="2025-01-11T11:31:08-05:00" level=debug msg="Cached value indicated that overlay is supported" time="2025-01-11T11:31:08-05:00" level=debug msg="Cached value indicated that metacopy is not being used" time="2025-01-11T11:31:08-05:00" level=debug msg="Cached value indicated that native-diff is usable" time="2025-01-11T11:31:08-05:00" level=debug msg="backingFs=xfs, projectQuotaSupported=false, useNativeDiff=true, usingMetacopy=false" time="2025-01-11T11:31:08-05:00" level=debug msg="Initializing event backend file" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime ocijail initialization failed: no valid executable found for OCI runtime ocijail: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime crun-vm initialization failed: no valid executable found for OCI runtime crun-vm: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime youki initialization failed: no valid executable found for OCI runtime youki: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime kata initialization failed: no valid executable found for OCI runtime kata: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime runsc initialization failed: no valid executable found for OCI runtime runsc: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime krun initialization failed: no valid executable found for OCI runtime krun: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime crun-wasm initialization failed: no valid executable found for OCI runtime crun-wasm: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime runc initialization failed: no valid executable found for OCI runtime runc: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime runj initialization failed: no valid executable found for OCI runtime runj: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Using OCI runtime \"/usr/bin/crun\"" time="2025-01-11T11:31:08-05:00" level=info msg="Setting parallel job count to 7" time="2025-01-11T11:31:08-05:00" level=debug msg="Successfully loaded 1 networks" time="2025-01-11T11:31:08-05:00" level=debug msg="found free device name podman1" time="2025-01-11T11:31:08-05:00" level=debug msg="found free ipv4 network subnet 10.89.0.0/24" time="2025-01-11T11:31:08-05:00" level=debug msg="Looking up image \"localhost/podman-pause:5.3.1-1733097600\" in local containers storage" time="2025-01-11T11:31:08-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:08-05:00" level=debug msg="Trying \"localhost/podman-pause:5.3.1-1733097600\" ..." time="2025-01-11T11:31:08-05:00" level=debug msg="reference \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]localhost/podman-pause:5.3.1-1733097600\" does not resolve to an image ID" time="2025-01-11T11:31:08-05:00" level=debug msg="Trying \"localhost/podman-pause:5.3.1-1733097600\" ..." time="2025-01-11T11:31:08-05:00" level=debug msg="reference \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]localhost/podman-pause:5.3.1-1733097600\" does not resolve to an image ID" time="2025-01-11T11:31:08-05:00" level=debug msg="Trying \"localhost/podman-pause:5.3.1-1733097600\" ..." time="2025-01-11T11:31:08-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:08-05:00" level=debug msg="FROM \"scratch\"" time="2025-01-11T11:31:08-05:00" level=debug msg="Cached value indicated that idmapped mounts for overlay are not supported" time="2025-01-11T11:31:08-05:00" level=debug msg="Check for idmapped mounts support " time="2025-01-11T11:31:08-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:08-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:09-05:00" level=debug msg="overlay: test mount indicated that volatile is being used" time="2025-01-11T11:31:09-05:00" level=debug msg="overlay: mount_data=lowerdir=/home/podman_basic_user/.local/share/containers/storage/overlay/ce5d425d1a334b7a751d21348811c909114800550f7ebdee36d06daf3bf588d0/empty,upperdir=/home/podman_basic_user/.local/share/containers/storage/overlay/ce5d425d1a334b7a751d21348811c909114800550f7ebdee36d06daf3bf588d0/diff,workdir=/home/podman_basic_user/.local/share/containers/storage/overlay/ce5d425d1a334b7a751d21348811c909114800550f7ebdee36d06daf3bf588d0/work,userxattr,volatile,context=\"system_u:object_r:container_file_t:s0:c681,c972\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Container ID: 108a4b73ea294c92e350e6914ba70ae43bc510d0757a34527239170a2fe2fa22" time="2025-01-11T11:31:09-05:00" level=debug msg="Parsed Step: {Env:[PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin] Command:copy Args:[/usr/libexec/podman/catatonit /catatonit] Flags:[] Attrs:map[] Message:COPY /usr/libexec/podman/catatonit /catatonit Heredocs:[] Original:COPY /usr/libexec/podman/catatonit /catatonit}" time="2025-01-11T11:31:09-05:00" level=debug msg="COPY []string(nil), imagebuilder.Copy{FromFS:false, From:\"\", Src:[]string{\"/usr/libexec/podman/catatonit\"}, Dest:\"/catatonit\", Download:false, Chown:\"\", Chmod:\"\", Checksum:\"\", Files:[]imagebuilder.File(nil), KeepGitDir:false, Link:false, Parents:false, Excludes:[]string(nil)}" time="2025-01-11T11:31:09-05:00" level=debug msg="EnsureContainerPath \"/\" (owner \"\", mode 0) in \"108a4b73ea294c92e350e6914ba70ae43bc510d0757a34527239170a2fe2fa22\"" time="2025-01-11T11:31:09-05:00" level=debug msg="added content file:872045bc026632654869cebc14f91fe531969760bf26925e889d9b70527c8b67" time="2025-01-11T11:31:09-05:00" level=debug msg="Parsed Step: {Env:[PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin] Command:entrypoint Args:[/catatonit -P] Flags:[] Attrs:map[json:true] Message:ENTRYPOINT /catatonit -P Heredocs:[] Original:ENTRYPOINT [\"/catatonit\", \"-P\"]}" time="2025-01-11T11:31:09-05:00" level=debug msg="EnsureContainerPath \"/\" (owner \"\", mode 0) in \"108a4b73ea294c92e350e6914ba70ae43bc510d0757a34527239170a2fe2fa22\"" time="2025-01-11T11:31:09-05:00" level=debug msg="COMMIT localhost/podman-pause:5.3.1-1733097600" time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]localhost/podman-pause:5.3.1-1733097600\"" time="2025-01-11T11:31:09-05:00" level=debug msg="COMMIT \"containers-storage:[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]localhost/podman-pause:5.3.1-1733097600\"" time="2025-01-11T11:31:09-05:00" level=debug msg="committing image with reference \"containers-storage:[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]localhost/podman-pause:5.3.1-1733097600\" is allowed by policy" time="2025-01-11T11:31:09-05:00" level=debug msg="layer list: [\"ce5d425d1a334b7a751d21348811c909114800550f7ebdee36d06daf3bf588d0\"]" time="2025-01-11T11:31:09-05:00" level=debug msg="using \"/var/tmp/buildah4034309036\" to hold temporary data" time="2025-01-11T11:31:09-05:00" level=debug msg="Tar with options on /home/podman_basic_user/.local/share/containers/storage/overlay/ce5d425d1a334b7a751d21348811c909114800550f7ebdee36d06daf3bf588d0/diff" time="2025-01-11T11:31:09-05:00" level=debug msg="layer \"ce5d425d1a334b7a751d21348811c909114800550f7ebdee36d06daf3bf588d0\" size is 699392 bytes, uncompressed digest sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6, possibly-compressed digest sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6" time="2025-01-11T11:31:09-05:00" level=debug msg="OCIv1 config = {\"created\":\"2025-01-11T16:31:09.16093935Z\",\"architecture\":\"amd64\",\"os\":\"linux\",\"config\":{\"Env\":[\"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin\"],\"Entrypoint\":[\"/catatonit\",\"-P\"],\"WorkingDir\":\"/\",\"Labels\":{\"io.buildah.version\":\"1.38.0\"}},\"rootfs\":{\"type\":\"layers\",\"diff_ids\":[\"sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6\"]},\"history\":[{\"created\":\"2025-01-11T16:31:09.133319185Z\",\"created_by\":\"/bin/sh -c #(nop) COPY file:872045bc026632654869cebc14f91fe531969760bf26925e889d9b70527c8b67 in /catatonit \",\"empty_layer\":true},{\"created\":\"2025-01-11T16:31:09.164063409Z\",\"created_by\":\"/bin/sh -c #(nop) ENTRYPOINT [\\\"/catatonit\\\", \\\"-P\\\"]\"}]}" time="2025-01-11T11:31:09-05:00" level=debug msg="OCIv1 manifest = {\"schemaVersion\":2,\"mediaType\":\"application/vnd.oci.image.manifest.v1+json\",\"config\":{\"mediaType\":\"application/vnd.oci.image.config.v1+json\",\"digest\":\"sha256:5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\",\"size\":684},\"layers\":[{\"mediaType\":\"application/vnd.oci.image.layer.v1.tar\",\"digest\":\"sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6\",\"size\":699392}],\"annotations\":{\"org.opencontainers.image.base.digest\":\"\",\"org.opencontainers.image.base.name\":\"\"}}" time="2025-01-11T11:31:09-05:00" level=debug msg="Docker v2s2 config = {\"created\":\"2025-01-11T16:31:09.16093935Z\",\"container\":\"108a4b73ea294c92e350e6914ba70ae43bc510d0757a34527239170a2fe2fa22\",\"container_config\":{\"Hostname\":\"\",\"Domainname\":\"\",\"User\":\"\",\"AttachStdin\":false,\"AttachStdout\":false,\"AttachStderr\":false,\"Tty\":false,\"OpenStdin\":false,\"StdinOnce\":false,\"Env\":[\"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin\"],\"Cmd\":null,\"Image\":\"\",\"Volumes\":{},\"WorkingDir\":\"/\",\"Entrypoint\":[\"/catatonit\",\"-P\"],\"OnBuild\":[],\"Labels\":{\"io.buildah.version\":\"1.38.0\"}},\"config\":{\"Hostname\":\"\",\"Domainname\":\"\",\"User\":\"\",\"AttachStdin\":false,\"AttachStdout\":false,\"AttachStderr\":false,\"Tty\":false,\"OpenStdin\":false,\"StdinOnce\":false,\"Env\":[\"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin\"],\"Cmd\":null,\"Image\":\"\",\"Volumes\":{},\"WorkingDir\":\"/\",\"Entrypoint\":[\"/catatonit\",\"-P\"],\"OnBuild\":[],\"Labels\":{\"io.buildah.version\":\"1.38.0\"}},\"architecture\":\"amd64\",\"os\":\"linux\",\"rootfs\":{\"type\":\"layers\",\"diff_ids\":[\"sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6\"]},\"history\":[{\"created\":\"2025-01-11T16:31:09.133319185Z\",\"created_by\":\"/bin/sh -c #(nop) COPY file:872045bc026632654869cebc14f91fe531969760bf26925e889d9b70527c8b67 in /catatonit \",\"empty_layer\":true},{\"created\":\"2025-01-11T16:31:09.164063409Z\",\"created_by\":\"/bin/sh -c #(nop) ENTRYPOINT [\\\"/catatonit\\\", \\\"-P\\\"]\"}]}" time="2025-01-11T11:31:09-05:00" level=debug msg="Docker v2s2 manifest = {\"schemaVersion\":2,\"mediaType\":\"application/vnd.docker.distribution.manifest.v2+json\",\"config\":{\"mediaType\":\"application/vnd.docker.container.image.v1+json\",\"size\":1347,\"digest\":\"sha256:3d5363d5b8025414d8363b8279ab1974a53f63e86ef685536caca8cc6f1f8c3a\"},\"layers\":[{\"mediaType\":\"application/vnd.docker.image.rootfs.diff.tar\",\"size\":699392,\"digest\":\"sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6\"}]}" time="2025-01-11T11:31:09-05:00" level=debug msg="Using SQLite blob info cache at /home/podman_basic_user/.local/share/containers/cache/blob-info-cache-v1.sqlite" time="2025-01-11T11:31:09-05:00" level=debug msg="IsRunningImageAllowed for image containers-storage:" time="2025-01-11T11:31:09-05:00" level=debug msg=" Using transport \"containers-storage\" policy section \"\"" time="2025-01-11T11:31:09-05:00" level=debug msg=" Requirement 0: allowed" time="2025-01-11T11:31:09-05:00" level=debug msg="Overall: allowed" time="2025-01-11T11:31:09-05:00" level=debug msg="start reading config" time="2025-01-11T11:31:09-05:00" level=debug msg="finished reading config" time="2025-01-11T11:31:09-05:00" level=debug msg="Manifest has MIME type application/vnd.oci.image.manifest.v1+json, ordered candidate list [application/vnd.oci.image.manifest.v1+json, application/vnd.docker.distribution.manifest.v2+json, application/vnd.docker.distribution.manifest.v1+prettyjws, application/vnd.docker.distribution.manifest.v1+json]" time="2025-01-11T11:31:09-05:00" level=debug msg="... will first try using the original manifest unmodified" time="2025-01-11T11:31:09-05:00" level=debug msg="Checking if we can reuse blob sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6: general substitution = true, compression for MIME type \"application/vnd.oci.image.layer.v1.tar\" = true" time="2025-01-11T11:31:09-05:00" level=debug msg="reading layer \"sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6\"" time="2025-01-11T11:31:09-05:00" level=debug msg="No compression detected" time="2025-01-11T11:31:09-05:00" level=debug msg="Using original blob without modification" time="2025-01-11T11:31:09-05:00" level=debug msg="Applying tar in /home/podman_basic_user/.local/share/containers/storage/overlay/83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6/diff" time="2025-01-11T11:31:09-05:00" level=debug msg="finished reading layer \"sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6\"" time="2025-01-11T11:31:09-05:00" level=debug msg="No compression detected" time="2025-01-11T11:31:09-05:00" level=debug msg="Compression change for blob sha256:5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49 (\"application/vnd.oci.image.config.v1+json\") not supported" time="2025-01-11T11:31:09-05:00" level=debug msg="Using original blob without modification" time="2025-01-11T11:31:09-05:00" level=debug msg="setting image creation date to 2025-01-11 16:31:09.16093935 +0000 UTC" time="2025-01-11T11:31:09-05:00" level=debug msg="created new image ID \"5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\" with metadata \"{}\"" time="2025-01-11T11:31:09-05:00" level=debug msg="added name \"localhost/podman-pause:5.3.1-1733097600\" to image \"5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\"" time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]localhost/podman-pause:5.3.1-1733097600\"" time="2025-01-11T11:31:09-05:00" level=debug msg="printing final image id \"5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Pod using bridge network mode" time="2025-01-11T11:31:09-05:00" level=debug msg="Created cgroup path user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice for parent user.slice and name libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da" time="2025-01-11T11:31:09-05:00" level=debug msg="Created cgroup user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice" time="2025-01-11T11:31:09-05:00" level=debug msg="Got pod cgroup as user.slice/user-3001.slice/user@3001.service/user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice" time="2025-01-11T11:31:09-05:00" level=debug msg="Looking up image \"localhost/podman-pause:5.3.1-1733097600\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:09-05:00" level=debug msg="Trying \"localhost/podman-pause:5.3.1-1733097600\" ..." time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"localhost/podman-pause:5.3.1-1733097600\" as \"localhost/podman-pause:5.3.1-1733097600\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"localhost/podman-pause:5.3.1-1733097600\" as \"localhost/podman-pause:5.3.1-1733097600\" in local containers storage ([overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49)" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49" time="2025-01-11T11:31:09-05:00" level=debug msg="using systemd mode: false" time="2025-01-11T11:31:09-05:00" level=debug msg="setting container name cec703044d41-infra" time="2025-01-11T11:31:09-05:00" level=debug msg="Loading seccomp profile from \"/usr/share/containers/seccomp.json\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Successfully loaded network podman-default-kube-network: &{podman-default-kube-network fb42127ceae097e95afb4c22c535aaa88e07cb9147b8dd0d6c01dc2c233a4302 bridge podman1 2025-01-11 11:31:08.992158363 -0500 EST [{{{10.89.0.0 ffffff00}} 10.89.0.1 }] [] false false true [] map[] map[] map[driver:host-local]}" time="2025-01-11T11:31:09-05:00" level=debug msg="Successfully loaded 2 networks" time="2025-01-11T11:31:09-05:00" level=debug msg="Allocated lock 1 for container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Created container \"6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Container \"6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e\" has work directory \"/home/podman_basic_user/.local/share/containers/storage/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Container \"6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e\" has run directory \"/run/user/3001/containers/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Looking up image \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:09-05:00" level=debug msg="Trying \"quay.io/libpod/testimage:20210610\" ..." time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage ([overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f)" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Pulling image quay.io/libpod/testimage:20210610 (policy: missing)" time="2025-01-11T11:31:09-05:00" level=debug msg="Looking up image \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:09-05:00" level=debug msg="Trying \"quay.io/libpod/testimage:20210610\" ..." time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage ([overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f)" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Looking up image \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:09-05:00" level=debug msg="Trying \"quay.io/libpod/testimage:20210610\" ..." time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage ([overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f)" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Looking up image \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:09-05:00" level=debug msg="Trying \"quay.io/libpod/testimage:20210610\" ..." time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage ([overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f)" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f" time="2025-01-11T11:31:09-05:00" level=debug msg="using systemd mode: false" time="2025-01-11T11:31:09-05:00" level=debug msg="adding container to pod httpd1" time="2025-01-11T11:31:09-05:00" level=debug msg="setting container name httpd1-httpd1" time="2025-01-11T11:31:09-05:00" level=debug msg="Loading seccomp profile from \"/usr/share/containers/seccomp.json\"" time="2025-01-11T11:31:09-05:00" level=info msg="Sysctl net.ipv4.ping_group_range=0 0 ignored in containers.conf, since Network Namespace set to host" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding mount /proc" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding mount /dev" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding mount /dev/pts" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding mount /dev/mqueue" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding mount /sys" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding mount /sys/fs/cgroup" time="2025-01-11T11:31:09-05:00" level=debug msg="Allocated lock 2 for container 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Created container \"69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Container \"69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de\" has work directory \"/home/podman_basic_user/.local/share/containers/storage/overlay-containers/69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de/userdata\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Container \"69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de\" has run directory \"/run/user/3001/containers/overlay-containers/69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de/userdata\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Strongconnecting node 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" time="2025-01-11T11:31:09-05:00" level=debug msg="Pushed 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e onto stack" time="2025-01-11T11:31:09-05:00" level=debug msg="Finishing node 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e. Popped 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e off stack" time="2025-01-11T11:31:09-05:00" level=debug msg="Strongconnecting node 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de" time="2025-01-11T11:31:09-05:00" level=debug msg="Pushed 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de onto stack" time="2025-01-11T11:31:09-05:00" level=debug msg="Finishing node 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de. Popped 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de off stack" time="2025-01-11T11:31:09-05:00" level=debug msg="overlay: mount_data=lowerdir=/home/podman_basic_user/.local/share/containers/storage/overlay/l/VY45JJLRBH5N22CXKVWPXHFS6U,upperdir=/home/podman_basic_user/.local/share/containers/storage/overlay/438cf6e5d85031b39f6e2286fb90558bcb520d90dfcaf76539b1082e932b7376/diff,workdir=/home/podman_basic_user/.local/share/containers/storage/overlay/438cf6e5d85031b39f6e2286fb90558bcb520d90dfcaf76539b1082e932b7376/work,userxattr,context=\"system_u:object_r:container_file_t:s0:c471,c842\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Made network namespace at /run/user/3001/netns/netns-1ba1aabb-9ce7-ed09-ced0-c13ed359f037 for container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" time="2025-01-11T11:31:09-05:00" level=debug msg="Mounted container \"6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e\" at \"/home/podman_basic_user/.local/share/containers/storage/overlay/438cf6e5d85031b39f6e2286fb90558bcb520d90dfcaf76539b1082e932b7376/merged\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Created root filesystem for container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e at /home/podman_basic_user/.local/share/containers/storage/overlay/438cf6e5d85031b39f6e2286fb90558bcb520d90dfcaf76539b1082e932b7376/merged" time="2025-01-11T11:31:09-05:00" level=debug msg="Creating rootless network namespace at \"/run/user/3001/containers/networks/rootless-netns/rootless-netns\"" time="2025-01-11T11:31:09-05:00" level=debug msg="pasta arguments: --config-net --pid /run/user/3001/containers/networks/rootless-netns/rootless-netns-conn.pid --dns-forward 169.254.1.1 -t none -u none -T none -U none --no-map-gw --quiet --netns /run/user/3001/containers/networks/rootless-netns/rootless-netns --map-guest-addr 169.254.1.2" time="2025-01-11T11:31:09-05:00" level=debug msg="The path of /etc/resolv.conf in the mount ns is \"/etc/resolv.conf\"" [DEBUG netavark::network::validation] Validating network namespace... [DEBUG netavark::commands::setup] Setting up... [INFO netavark::firewall] Using nftables firewall driver [DEBUG netavark::network::bridge] Setup network podman-default-kube-network [DEBUG netavark::network::bridge] Container interface name: eth0 with IP addresses [10.89.0.2/24] [DEBUG netavark::network::bridge] Bridge name: podman1 with IP addresses [10.89.0.1/24] [DEBUG netavark::network::core_utils] Setting sysctl value for net.ipv4.ip_forward to 1 [DEBUG netavark::network::core_utils] Setting sysctl value for /proc/sys/net/ipv4/conf/podman1/rp_filter to 2 [DEBUG netavark::network::core_utils] Setting sysctl value for /proc/sys/net/ipv6/conf/eth0/autoconf to 0 [DEBUG netavark::network::core_utils] Setting sysctl value for /proc/sys/net/ipv4/conf/eth0/arp_notify to 1 [DEBUG netavark::network::core_utils] Setting sysctl value for /proc/sys/net/ipv4/conf/eth0/rp_filter to 2 [INFO netavark::network::netlink] Adding route (dest: 0.0.0.0/0 ,gw: 10.89.0.1, metric 100) [INFO netavark::firewall::nft] Creating container chain nv_fb42127c_10_89_0_0_nm24 [DEBUG netavark::network::core_utils] Setting sysctl value for net.ipv4.conf.podman1.route_localnet to 1 [DEBUG netavark::dns::aardvark] Spawning aardvark server [DEBUG netavark::dns::aardvark] start aardvark-dns: ["systemd-run", "-q", "--scope", "--user", "/usr/libexec/podman/aardvark-dns", "--config", "/run/user/3001/containers/networks/aardvark-dns", "-p", "53", "run"] [DEBUG netavark::commands::setup] { "podman-default-kube-network": StatusBlock { dns_search_domains: Some( [ "dns.podman", ], ), dns_server_ips: Some( [ 10.89.0.1, ], ), interfaces: Some( { "eth0": NetInterface { mac_address: "ba:4e:01:fd:9a:cc", subnets: Some( [ NetAddress { gateway: Some( 10.89.0.1, ), ipnet: 10.89.0.2/24, }, ], ), }, }, ), }, } [DEBUG netavark::commands::setup] Setup complete time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport: time=\"2025-01-11T11:31:09-05:00\" level=info msg=\"Starting parent driver\"\n" time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport: time=\"2025-01-11T11:31:09-05:00\" level=info msg=\"opaque=map[builtin.readypipepath:/run/user/3001/libpod/tmp/rootlessport1778471301/.bp-ready.pipe builtin.socketpath:/run/user/3001/libpod/tmp/rootlessport1778471301/.bp.sock]\"\n" time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport: time=\"2025-01-11T11:31:09-05:00\" level=info msg=\"Starting child driver in child netns (\\\"/proc/self/exe\\\" [rootlessport-child])\"\n" time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport: time=\"2025-01-11T11:31:09-05:00\" level=info msg=\"Waiting for initComplete\"\n" time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport: time=\"2025-01-11T11:31:09-05:00\" level=info msg=\"initComplete is closed; parent and child established the communication channel\"\ntime=\"2025-01-11T11:31:09-05:00\" level=info msg=\"Exposing ports [{ 80 15001 1 tcp}]\"\n" time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport: time=\"2025-01-11T11:31:09-05:00\" level=info msg=Ready\n" time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport is ready" time="2025-01-11T11:31:09-05:00" level=debug msg="/proc/sys/crypto/fips_enabled does not contain '1', not adding FIPS mode bind mounts" time="2025-01-11T11:31:09-05:00" level=debug msg="Setting Cgroups for container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e to user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice:libpod:6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" time="2025-01-11T11:31:09-05:00" level=debug msg="reading hooks from /usr/share/containers/oci/hooks.d" time="2025-01-11T11:31:09-05:00" level=debug msg="Workdir \"/\" resolved to host path \"/home/podman_basic_user/.local/share/containers/storage/overlay/438cf6e5d85031b39f6e2286fb90558bcb520d90dfcaf76539b1082e932b7376/merged\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Created OCI spec for container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e at /home/podman_basic_user/.local/share/containers/storage/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata/config.json" time="2025-01-11T11:31:09-05:00" level=debug msg="Created cgroup path user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice for parent user.slice and name libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da" time="2025-01-11T11:31:09-05:00" level=debug msg="Created cgroup user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice" time="2025-01-11T11:31:09-05:00" level=debug msg="Got pod cgroup as user.slice/user-3001.slice/user@3001.service/user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice" time="2025-01-11T11:31:09-05:00" level=debug msg="/usr/bin/conmon messages will be logged to syslog" time="2025-01-11T11:31:09-05:00" level=debug msg="running conmon: /usr/bin/conmon" args="[--api-version 1 -c 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e -u 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e -r /usr/bin/crun -b /home/podman_basic_user/.local/share/containers/storage/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata -p /run/user/3001/containers/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata/pidfile -n cec703044d41-infra --exit-dir /run/user/3001/libpod/tmp/exits --persist-dir /run/user/3001/libpod/tmp/persist/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e --full-attach -s -l k8s-file:/home/podman_basic_user/.local/share/containers/storage/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata/ctr.log --log-level debug --syslog --conmon-pidfile /run/user/3001/containers/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata/conmon.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /home/podman_basic_user/.local/share/containers/storage --exit-command-arg --runroot --exit-command-arg /run/user/3001/containers --exit-command-arg --log-level --exit-command-arg debug --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/user/3001/libpod/tmp --exit-command-arg --network-config-dir --exit-command-arg --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /home/podman_basic_user/.local/share/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --events-backend --exit-command-arg file --exit-command-arg --syslog --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e]" time="2025-01-11T11:31:09-05:00" level=info msg="Running conmon under slice user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice and unitName libpod-conmon-6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e.scope" [conmon:d]: failed to write to /proc/self/oom_score_adj: Permission denied time="2025-01-11T11:31:09-05:00" level=debug msg="Received: 24217" time="2025-01-11T11:31:09-05:00" level=info msg="Got Conmon PID as 24215" time="2025-01-11T11:31:09-05:00" level=debug msg="Created container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e in OCI runtime" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding nameserver(s) from network status of '[\"10.89.0.1\"]'" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding search domain(s) from network status of '[\"dns.podman\"]'" time="2025-01-11T11:31:09-05:00" level=debug msg="Starting container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e with command [/catatonit -P]" time="2025-01-11T11:31:09-05:00" level=debug msg="Started container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" time="2025-01-11T11:31:09-05:00" level=debug msg="overlay: mount_data=lowerdir=/home/podman_basic_user/.local/share/containers/storage/overlay/l/H6MLQJZURTWK5UKPYM3NBCTJTC,upperdir=/home/podman_basic_user/.local/share/containers/storage/overlay/2fa690399eece911fb5509ec500fd7d2ca654be99b65c5d16219b201e4a935b1/diff,workdir=/home/podman_basic_user/.local/share/containers/storage/overlay/2fa690399eece911fb5509ec500fd7d2ca654be99b65c5d16219b201e4a935b1/work,userxattr,context=\"system_u:object_r:container_file_t:s0:c471,c842\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Mounted container \"69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de\" at \"/home/podman_basic_user/.local/share/containers/storage/overlay/2fa690399eece911fb5509ec500fd7d2ca654be99b65c5d16219b201e4a935b1/merged\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Created root filesystem for container 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de at /home/podman_basic_user/.local/share/containers/storage/overlay/2fa690399eece911fb5509ec500fd7d2ca654be99b65c5d16219b201e4a935b1/merged" time="2025-01-11T11:31:09-05:00" level=debug msg="/proc/sys/crypto/fips_enabled does not contain '1', not adding FIPS mode bind mounts" time="2025-01-11T11:31:09-05:00" level=debug msg="Cleaning up container 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de" time="2025-01-11T11:31:09-05:00" level=debug msg="Unmounted container \"69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de\"" starting container 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de: cannot get namespace path unless container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e is running: container is stopped Error: failed to start 1 containers time="2025-01-11T11:31:09-05:00" level=debug msg="Shutting down engines" time="2025-01-11T11:31:09-05:00" level=info msg="Received shutdown.Stop(), terminating!" PID=24117 time="2025-01-11T11:31:09-05:00" level=debug msg="Adding parallel job to stop container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding parallel job to stop container 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de" time="2025-01-11T11:31:09-05:00" level=debug msg="Stopping ctr 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de (timeout 10)" time="2025-01-11T11:31:09-05:00" level=debug msg="Stopping ctr 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e (timeout 10)" time="2025-01-11T11:31:09-05:00" level=debug msg="Removing pod cgroup user.slice/user-3001.slice/user@3001.service/user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice" TASK [Dump journal] ************************************************************ task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:447 Saturday 11 January 2025 11:31:09 -0500 (0:00:01.370) 0:02:34.854 ****** fatal: [managed-node2]: FAILED! => { "changed": false, "cmd": [ "journalctl", "-ex" ], "delta": "0:00:00.032257", "end": "2025-01-11 11:31:10.226923", "failed_when_result": true, "rc": 0, "start": "2025-01-11 11:31:10.194666" } STDOUT: Jan 11 11:24:47 localhost systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 11 11:24:47 localhost systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 11 11:24:47 localhost systemd-journald[528]: Journal started ░░ Subject: The journal has been started ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The system journal process has started up, opened the journal ░░ files for writing and is now ready to process requests. Jan 11 11:24:47 localhost systemd-journald[528]: Runtime Journal (/run/log/journal/ec22cc6fef820591f15f8e72441572da) is 8M, max 70.5M, 62.5M free. ░░ Subject: Disk space used by the journal ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ Runtime Journal (/run/log/journal/ec22cc6fef820591f15f8e72441572da) is currently using 8M. ░░ Maximum allowed usage is set to 70.5M. ░░ Leaving at least 35.2M free (of currently available 689.3M of disk space). ░░ Enforced usage limit is thus 70.5M, of which 62.5M are still available. ░░ ░░ The limits controlling how much disk space is used by the journal may ░░ be configured with SystemMaxUse=, SystemKeepFree=, SystemMaxFileSize=, ░░ RuntimeMaxUse=, RuntimeKeepFree=, RuntimeMaxFileSize= settings in ░░ /etc/systemd/journald.conf. See journald.conf(5) for details. Jan 11 11:24:45 localhost systemd-rc-local-generator[486]: /etc/rc.d/rc.local is not marked executable, skipping. Jan 11 11:24:46 localhost systemd[1]: Queued start job for default target multi-user.target. Jan 11 11:24:47 localhost systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 11 11:24:46 localhost systemd[1]: systemd-journald.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-journald.service has successfully entered the 'dead' state. Jan 11 11:24:47 localhost systemd[1]: Started systemd-journald.service - Journal Service. Jan 11 11:24:47 localhost systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. ░░ Subject: A start job for unit systemd-network-generator.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-network-generator.service has finished successfully. ░░ ░░ The job identifier is 196. Jan 11 11:24:47 localhost systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. ░░ Subject: A start job for unit systemd-remount-fs.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-remount-fs.service has finished successfully. ░░ ░░ The job identifier is 144. Jan 11 11:24:47 localhost systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. ░░ Subject: A start job for unit systemd-sysctl.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-sysctl.service has finished successfully. ░░ ░░ The job identifier is 173. Jan 11 11:24:47 localhost systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. ░░ Subject: A start job for unit systemd-udev-load-credentials.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-udev-load-credentials.service has finished successfully. ░░ ░░ The job identifier is 137. Jan 11 11:24:47 localhost systemd[1]: systemd-hwdb-update.service - Rebuild Hardware Database was skipped because of an unmet condition check (ConditionNeedsUpdate=/etc). ░░ Subject: A start job for unit systemd-hwdb-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hwdb-update.service has finished successfully. ░░ ░░ The job identifier is 141. Jan 11 11:24:47 localhost systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... ░░ Subject: A start job for unit systemd-journal-flush.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-journal-flush.service has begun execution. ░░ ░░ The job identifier is 192. Jan 11 11:24:47 localhost systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). ░░ Subject: A start job for unit systemd-pstore.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-pstore.service has finished successfully. ░░ ░░ The job identifier is 169. Jan 11 11:24:47 localhost systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... ░░ Subject: A start job for unit systemd-random-seed.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-random-seed.service has begun execution. ░░ ░░ The job identifier is 185. Jan 11 11:24:47 localhost systemd[1]: systemd-tpm2-setup.service - TPM SRK Setup was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ░░ Subject: A start job for unit systemd-tpm2-setup.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tpm2-setup.service has finished successfully. ░░ ░░ The job identifier is 153. Jan 11 11:24:47 localhost systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. ░░ Subject: A start job for unit systemd-random-seed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-random-seed.service has finished successfully. ░░ ░░ The job identifier is 185. Jan 11 11:24:47 localhost systemd-journald[528]: Runtime Journal (/run/log/journal/ec22cc6fef820591f15f8e72441572da) is 8M, max 70.5M, 62.5M free. ░░ Subject: Disk space used by the journal ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ Runtime Journal (/run/log/journal/ec22cc6fef820591f15f8e72441572da) is currently using 8M. ░░ Maximum allowed usage is set to 70.5M. ░░ Leaving at least 35.2M free (of currently available 689.2M of disk space). ░░ Enforced usage limit is thus 70.5M, of which 62.5M are still available. ░░ ░░ The limits controlling how much disk space is used by the journal may ░░ be configured with SystemMaxUse=, SystemKeepFree=, SystemMaxFileSize=, ░░ RuntimeMaxUse=, RuntimeKeepFree=, RuntimeMaxFileSize= settings in ░░ /etc/systemd/journald.conf. See journald.conf(5) for details. Jan 11 11:24:47 localhost systemd-journald[528]: Received client request to flush runtime journal. Jan 11 11:24:47 localhost systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. ░░ Subject: A start job for unit systemd-journal-flush.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-journal-flush.service has finished successfully. ░░ ░░ The job identifier is 192. Jan 11 11:24:47 localhost systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. ░░ Subject: A start job for unit systemd-udev-trigger.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-udev-trigger.service has finished successfully. ░░ ░░ The job identifier is 132. Jan 11 11:24:47 localhost systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. ░░ Subject: A start job for unit systemd-tmpfiles-setup-dev-early.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-setup-dev-early.service has finished successfully. ░░ ░░ The job identifier is 166. Jan 11 11:24:47 localhost systemd[1]: systemd-sysusers.service - Create System Users was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit systemd-sysusers.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-sysusers.service has finished successfully. ░░ ░░ The job identifier is 181. Jan 11 11:24:47 localhost systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... ░░ Subject: A start job for unit systemd-tmpfiles-setup-dev.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-setup-dev.service has begun execution. ░░ ░░ The job identifier is 167. Jan 11 11:24:47 localhost systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. ░░ Subject: A start job for unit systemd-tmpfiles-setup-dev.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-setup-dev.service has finished successfully. ░░ ░░ The job identifier is 167. Jan 11 11:24:47 localhost systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... ░░ Subject: A start job for unit systemd-udevd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-udevd.service has begun execution. ░░ ░░ The job identifier is 133. Jan 11 11:24:47 localhost systemd-udevd[566]: Using default interface naming scheme 'rhel-10.0-beta'. Jan 11 11:24:47 localhost systemd[1]: Finished lvm2-monitor.service - Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling. ░░ Subject: A start job for unit lvm2-monitor.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit lvm2-monitor.service has finished successfully. ░░ ░░ The job identifier is 183. Jan 11 11:24:47 localhost systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. ░░ Subject: A start job for unit local-fs-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit local-fs-pre.target has finished successfully. ░░ ░░ The job identifier is 145. Jan 11 11:24:47 localhost systemd[1]: Reached target local-fs.target - Local File Systems. ░░ Subject: A start job for unit local-fs.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit local-fs.target has finished successfully. ░░ ░░ The job identifier is 143. Jan 11 11:24:47 localhost systemd[1]: Listening on systemd-bootctl.socket - Boot Entries Service Socket. ░░ Subject: A start job for unit systemd-bootctl.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-bootctl.socket has finished successfully. ░░ ░░ The job identifier is 222. Jan 11 11:24:47 localhost systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. ░░ Subject: A start job for unit systemd-sysext.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-sysext.socket has finished successfully. ░░ ░░ The job identifier is 221. Jan 11 11:24:47 localhost systemd[1]: selinux-autorelabel-mark.service - Mark the need to relabel after reboot was skipped because of an unmet condition check (ConditionSecurity=!selinux). ░░ Subject: A start job for unit selinux-autorelabel-mark.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit selinux-autorelabel-mark.service has finished successfully. ░░ ░░ The job identifier is 171. Jan 11 11:24:47 localhost systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit systemd-binfmt.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-binfmt.service has finished successfully. ░░ ░░ The job identifier is 182. Jan 11 11:24:47 localhost systemd[1]: systemd-boot-random-seed.service - Update Boot Loader Random Seed was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit systemd-boot-random-seed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-boot-random-seed.service has finished successfully. ░░ ░░ The job identifier is 168. Jan 11 11:24:47 localhost systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit systemd-confext.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-confext.service has finished successfully. ░░ ░░ The job identifier is 179. Jan 11 11:24:47 localhost systemd[1]: systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/ was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit systemd-sysext.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-sysext.service has finished successfully. ░░ ░░ The job identifier is 131. Jan 11 11:24:47 localhost systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... ░░ Subject: A start job for unit systemd-tmpfiles-setup.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-setup.service has begun execution. ░░ ░░ The job identifier is 124. Jan 11 11:24:47 localhost systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. ░░ Subject: A start job for unit systemd-tmpfiles-setup.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-setup.service has finished successfully. ░░ ░░ The job identifier is 124. Jan 11 11:24:47 localhost systemd[1]: Starting audit-rules.service - Load Audit Rules... ░░ Subject: A start job for unit audit-rules.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit audit-rules.service has begun execution. ░░ ░░ The job identifier is 237. Jan 11 11:24:47 localhost systemd[1]: ldconfig.service - Rebuild Dynamic Linker Cache was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit ldconfig.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit ldconfig.service has finished successfully. ░░ ░░ The job identifier is 146. Jan 11 11:24:47 localhost systemd[1]: Starting rpcbind.service - RPC Bind... ░░ Subject: A start job for unit rpcbind.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpcbind.service has begun execution. ░░ ░░ The job identifier is 238. Jan 11 11:24:47 localhost systemd[1]: systemd-firstboot.service - First Boot Wizard was skipped because of an unmet condition check (ConditionFirstBoot=yes). ░░ Subject: A start job for unit systemd-firstboot.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-firstboot.service has finished successfully. ░░ ░░ The job identifier is 151. Jan 11 11:24:47 localhost systemd[1]: first-boot-complete.target - First Boot Complete was skipped because of an unmet condition check (ConditionFirstBoot=yes). ░░ Subject: A start job for unit first-boot-complete.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit first-boot-complete.target has finished successfully. ░░ ░░ The job identifier is 152. Jan 11 11:24:47 localhost systemd[1]: systemd-journal-catalog-update.service - Rebuild Journal Catalog was skipped because of an unmet condition check (ConditionNeedsUpdate=/var). ░░ Subject: A start job for unit systemd-journal-catalog-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-journal-catalog-update.service has finished successfully. ░░ ░░ The job identifier is 199. Jan 11 11:24:47 localhost systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... ░░ Subject: A start job for unit systemd-machine-id-commit.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-machine-id-commit.service has begun execution. ░░ ░░ The job identifier is 178. Jan 11 11:24:47 localhost systemd[1]: systemd-update-done.service - Update is Completed was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit systemd-update-done.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-done.service has finished successfully. ░░ ░░ The job identifier is 193. Jan 11 11:24:47 localhost systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... ░░ Subject: A start job for unit sys-fs-fuse-connections.mount has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sys-fs-fuse-connections.mount has begun execution. ░░ ░░ The job identifier is 176. Jan 11 11:24:47 localhost systemd[1]: Mounting var-lib-nfs-rpc_pipefs.mount - RPC Pipe File System... ░░ Subject: A start job for unit var-lib-nfs-rpc_pipefs.mount has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit var-lib-nfs-rpc_pipefs.mount has begun execution. ░░ ░░ The job identifier is 247. Jan 11 11:24:47 localhost systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. ░░ Subject: A start job for unit sys-fs-fuse-connections.mount has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sys-fs-fuse-connections.mount has finished successfully. ░░ ░░ The job identifier is 176. Jan 11 11:24:48 localhost kernel: RPC: Registered named UNIX socket transport module. Jan 11 11:24:48 localhost kernel: RPC: Registered udp transport module. Jan 11 11:24:48 localhost kernel: RPC: Registered tcp transport module. Jan 11 11:24:48 localhost kernel: RPC: Registered tcp-with-tls transport module. Jan 11 11:24:48 localhost kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Jan 11 11:24:48 localhost systemd[1]: Mounted var-lib-nfs-rpc_pipefs.mount - RPC Pipe File System. ░░ Subject: A start job for unit var-lib-nfs-rpc_pipefs.mount has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit var-lib-nfs-rpc_pipefs.mount has finished successfully. ░░ ░░ The job identifier is 247. Jan 11 11:24:48 localhost systemd[1]: Reached target rpc_pipefs.target. ░░ Subject: A start job for unit rpc_pipefs.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc_pipefs.target has finished successfully. ░░ ░░ The job identifier is 246. Jan 11 11:24:48 localhost systemd[1]: etc-machine\x2did.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit etc-machine\x2did.mount has successfully entered the 'dead' state. Jan 11 11:24:48 localhost systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. ░░ Subject: A start job for unit systemd-machine-id-commit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-machine-id-commit.service has finished successfully. ░░ ░░ The job identifier is 178. Jan 11 11:24:48 localhost systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. ░░ Subject: A start job for unit systemd-udevd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-udevd.service has finished successfully. ░░ ░░ The job identifier is 133. Jan 11 11:24:48 localhost systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... ░░ Subject: A start job for unit modprobe@configfs.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@configfs.service has begun execution. ░░ ░░ The job identifier is 293. Jan 11 11:24:48 localhost systemd[1]: modprobe@configfs.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit modprobe@configfs.service has successfully entered the 'dead' state. Jan 11 11:24:48 localhost systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. ░░ Subject: A start job for unit modprobe@configfs.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@configfs.service has finished successfully. ░░ ░░ The job identifier is 293. Jan 11 11:24:48 localhost systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. ░░ Subject: A start job for unit dev-ttyS0.device has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dev-ttyS0.device has finished successfully. ░░ ░░ The job identifier is 230. Jan 11 11:24:48 localhost augenrules[571]: /sbin/augenrules: No change Jan 11 11:24:48 localhost augenrules[618]: No rules Jan 11 11:24:48 localhost augenrules[618]: enabled 0 Jan 11 11:24:48 localhost augenrules[618]: failure 1 Jan 11 11:24:48 localhost augenrules[618]: pid 0 Jan 11 11:24:48 localhost augenrules[618]: rate_limit 0 Jan 11 11:24:48 localhost augenrules[618]: backlog_limit 8192 Jan 11 11:24:48 localhost augenrules[618]: lost 0 Jan 11 11:24:48 localhost augenrules[618]: backlog 0 Jan 11 11:24:48 localhost augenrules[618]: backlog_wait_time 60000 Jan 11 11:24:48 localhost augenrules[618]: backlog_wait_time_actual 0 Jan 11 11:24:48 localhost augenrules[618]: enabled 0 Jan 11 11:24:48 localhost augenrules[618]: failure 1 Jan 11 11:24:48 localhost augenrules[618]: pid 0 Jan 11 11:24:48 localhost augenrules[618]: rate_limit 0 Jan 11 11:24:48 localhost augenrules[618]: backlog_limit 8192 Jan 11 11:24:48 localhost augenrules[618]: lost 0 Jan 11 11:24:48 localhost augenrules[618]: backlog 0 Jan 11 11:24:48 localhost augenrules[618]: backlog_wait_time 60000 Jan 11 11:24:48 localhost augenrules[618]: backlog_wait_time_actual 0 Jan 11 11:24:48 localhost augenrules[618]: enabled 0 Jan 11 11:24:48 localhost augenrules[618]: failure 1 Jan 11 11:24:48 localhost augenrules[618]: pid 0 Jan 11 11:24:48 localhost augenrules[618]: rate_limit 0 Jan 11 11:24:48 localhost augenrules[618]: backlog_limit 8192 Jan 11 11:24:48 localhost augenrules[618]: lost 0 Jan 11 11:24:48 localhost augenrules[618]: backlog 0 Jan 11 11:24:48 localhost augenrules[618]: backlog_wait_time 60000 Jan 11 11:24:48 localhost augenrules[618]: backlog_wait_time_actual 0 Jan 11 11:24:48 localhost systemd[1]: audit-rules.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit audit-rules.service has successfully entered the 'dead' state. Jan 11 11:24:48 localhost systemd[1]: Finished audit-rules.service - Load Audit Rules. ░░ Subject: A start job for unit audit-rules.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit audit-rules.service has finished successfully. ░░ ░░ The job identifier is 237. Jan 11 11:24:48 localhost systemd[1]: Started rpcbind.service - RPC Bind. ░░ Subject: A start job for unit rpcbind.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpcbind.service has finished successfully. ░░ ░░ The job identifier is 238. Jan 11 11:24:48 localhost systemd[1]: Starting auditd.service - Security Audit Logging Service... ░░ Subject: A start job for unit auditd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has begun execution. ░░ ░░ The job identifier is 236. Jan 11 11:24:48 localhost kernel: input: PC Speaker as /devices/platform/pcspkr/input/input5 Jan 11 11:24:48 localhost auditd[627]: No plugins found, not dispatching events Jan 11 11:24:48 localhost auditd[627]: Init complete, auditd 4.0 listening for events (startup state enable) Jan 11 11:24:48 localhost systemd[1]: Started auditd.service - Security Audit Logging Service. ░░ Subject: A start job for unit auditd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has finished successfully. ░░ ░░ The job identifier is 236. Jan 11 11:24:48 localhost systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... ░░ Subject: A start job for unit systemd-update-utmp.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp.service has begun execution. ░░ ░░ The job identifier is 241. Jan 11 11:24:48 localhost systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. ░░ Subject: A start job for unit systemd-update-utmp.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp.service has finished successfully. ░░ ░░ The job identifier is 241. Jan 11 11:24:48 localhost systemd[1]: Reached target sysinit.target - System Initialization. ░░ Subject: A start job for unit sysinit.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sysinit.target has finished successfully. ░░ ░░ The job identifier is 123. Jan 11 11:24:48 localhost systemd[1]: Started dnf-makecache.timer - dnf makecache --timer. ░░ Subject: A start job for unit dnf-makecache.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dnf-makecache.timer has finished successfully. ░░ ░░ The job identifier is 203. Jan 11 11:24:48 localhost systemd[1]: Started fstrim.timer - Discard unused filesystem blocks once a week. ░░ Subject: A start job for unit fstrim.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit fstrim.timer has finished successfully. ░░ ░░ The job identifier is 211. Jan 11 11:24:48 localhost systemd[1]: Started logrotate.timer - Daily rotation of log files. ░░ Subject: A start job for unit logrotate.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.timer has finished successfully. ░░ ░░ The job identifier is 210. Jan 11 11:24:48 localhost systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. ░░ Subject: A start job for unit systemd-tmpfiles-clean.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-clean.timer has finished successfully. ░░ ░░ The job identifier is 212. Jan 11 11:24:48 localhost systemd[1]: Reached target timers.target - Timer Units. ░░ Subject: A start job for unit timers.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit timers.target has finished successfully. ░░ ░░ The job identifier is 202. Jan 11 11:24:48 localhost kernel: cirrus 0000:00:02.0: vgaarb: deactivate vga console Jan 11 11:24:48 localhost kernel: Console: switching to colour dummy device 80x25 Jan 11 11:24:48 localhost kernel: [drm] Initialized cirrus 2.0.0 for 0000:00:02.0 on minor 0 Jan 11 11:24:48 localhost kernel: fbcon: cirrusdrmfb (fb0) is primary device Jan 11 11:24:48 localhost kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Jan 11 11:24:48 localhost kernel: Console: switching to colour frame buffer device 128x48 Jan 11 11:24:48 localhost kernel: cirrus 0000:00:02.0: [drm] fb0: cirrusdrmfb frame buffer device Jan 11 11:24:48 localhost systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. ░░ Subject: A start job for unit dbus.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus.socket has finished successfully. ░░ ░░ The job identifier is 208. Jan 11 11:24:48 localhost systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). ░░ Subject: A start job for unit sshd-unix-local.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-unix-local.socket has finished successfully. ░░ ░░ The job identifier is 219. Jan 11 11:24:48 localhost systemd[1]: Listening on sssd-kcm.socket - SSSD Kerberos Cache Manager responder socket. ░░ Subject: A start job for unit sssd-kcm.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sssd-kcm.socket has finished successfully. ░░ ░░ The job identifier is 223. Jan 11 11:24:48 localhost systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. ░░ Subject: A start job for unit systemd-hostnamed.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.socket has finished successfully. ░░ ░░ The job identifier is 220. Jan 11 11:24:48 localhost systemd[1]: Reached target sockets.target - Socket Units. ░░ Subject: A start job for unit sockets.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sockets.target has finished successfully. ░░ ░░ The job identifier is 213. Jan 11 11:24:48 localhost (udev-worker)[610]: Network interface NamePolicy= disabled on kernel command line. Jan 11 11:24:48 localhost systemd[1]: Starting dbus-broker.service - D-Bus System Message Bus... ░░ Subject: A start job for unit dbus-broker.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus-broker.service has begun execution. ░░ ░░ The job identifier is 207. Jan 11 11:24:48 localhost systemd[1]: systemd-pcrphase-sysinit.service - TPM PCR Barrier (Initialization) was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ░░ Subject: A start job for unit systemd-pcrphase-sysinit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-pcrphase-sysinit.service has finished successfully. ░░ ░░ The job identifier is 140. Jan 11 11:24:48 localhost systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... ░░ Subject: A start job for unit systemd-vconsole-setup.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-vconsole-setup.service has begun execution. ░░ ░░ The job identifier is 300. Jan 11 11:24:48 localhost systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-vconsole-setup.service has successfully entered the 'dead' state. Jan 11 11:24:48 localhost systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. ░░ Subject: A stop job for unit systemd-vconsole-setup.service has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit systemd-vconsole-setup.service has finished. ░░ ░░ The job identifier is 300 and the job result is done. Jan 11 11:24:48 localhost systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... ░░ Subject: A start job for unit systemd-vconsole-setup.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-vconsole-setup.service has begun execution. ░░ ░░ The job identifier is 300. Jan 11 11:24:48 localhost systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-vconsole-setup.service has successfully entered the 'dead' state. Jan 11 11:24:48 localhost systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. ░░ Subject: A stop job for unit systemd-vconsole-setup.service has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit systemd-vconsole-setup.service has finished. ░░ ░░ The job identifier is 300 and the job result is done. Jan 11 11:24:48 localhost systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... ░░ Subject: A start job for unit systemd-vconsole-setup.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-vconsole-setup.service has begun execution. ░░ ░░ The job identifier is 300. Jan 11 11:24:48 localhost kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 655360 ms ovfl timer Jan 11 11:24:48 localhost systemd[1]: Started dbus-broker.service - D-Bus System Message Bus. ░░ Subject: A start job for unit dbus-broker.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus-broker.service has finished successfully. ░░ ░░ The job identifier is 207. Jan 11 11:24:48 localhost systemd[1]: Reached target basic.target - Basic System. ░░ Subject: A start job for unit basic.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit basic.target has finished successfully. ░░ ░░ The job identifier is 122. Jan 11 11:24:48 localhost dbus-broker-launch[636]: Ready Jan 11 11:24:48 localhost systemd[1]: Starting chronyd.service - NTP client/server... ░░ Subject: A start job for unit chronyd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has begun execution. ░░ ░░ The job identifier is 265. Jan 11 11:24:48 localhost systemd[1]: Starting cloud-init-local.service - Initial cloud-init job (pre-networking)... ░░ Subject: A start job for unit cloud-init-local.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init-local.service has begun execution. ░░ ░░ The job identifier is 270. Jan 11 11:24:48 localhost systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... ░░ Subject: A start job for unit dracut-shutdown.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dracut-shutdown.service has begun execution. ░░ ░░ The job identifier is 142. Jan 11 11:24:48 localhost systemd[1]: Started irqbalance.service - irqbalance daemon. ░░ Subject: A start job for unit irqbalance.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit irqbalance.service has finished successfully. ░░ ░░ The job identifier is 278. Jan 11 11:24:48 localhost systemd[1]: Started rngd.service - Hardware RNG Entropy Gatherer Daemon. ░░ Subject: A start job for unit rngd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rngd.service has finished successfully. ░░ ░░ The job identifier is 261. Jan 11 11:24:48 localhost systemd[1]: ssh-host-keys-migration.service - Update OpenSSH host key permissions was skipped because of an unmet condition check (ConditionPathExists=!/var/lib/.ssh-host-keys-migration). ░░ Subject: A start job for unit ssh-host-keys-migration.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit ssh-host-keys-migration.service has finished successfully. ░░ ░░ The job identifier is 254. Jan 11 11:24:48 localhost systemd[1]: sshd-keygen@ecdsa.service - OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ecdsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ecdsa.service has finished successfully. ░░ ░░ The job identifier is 256. Jan 11 11:24:48 localhost systemd[1]: sshd-keygen@ed25519.service - OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ed25519.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ed25519.service has finished successfully. ░░ ░░ The job identifier is 258. Jan 11 11:24:48 localhost systemd[1]: sshd-keygen@rsa.service - OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@rsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@rsa.service has finished successfully. ░░ ░░ The job identifier is 259. Jan 11 11:24:48 localhost systemd[1]: Reached target sshd-keygen.target. ░░ Subject: A start job for unit sshd-keygen.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen.target has finished successfully. ░░ ░░ The job identifier is 255. Jan 11 11:24:48 localhost systemd[1]: sssd.service - System Security Services Daemon was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit sssd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sssd.service has finished successfully. ░░ ░░ The job identifier is 226. Jan 11 11:24:48 localhost systemd[1]: Reached target nss-user-lookup.target - User and Group Name Lookups. ░░ Subject: A start job for unit nss-user-lookup.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit nss-user-lookup.target has finished successfully. ░░ ░░ The job identifier is 227. Jan 11 11:24:48 localhost systemd[1]: Starting systemd-logind.service - User Login Management... ░░ Subject: A start job for unit systemd-logind.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-logind.service has begun execution. ░░ ░░ The job identifier is 262. Jan 11 11:24:48 localhost systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. ░░ Subject: A start job for unit systemd-vconsole-setup.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-vconsole-setup.service has finished successfully. ░░ ░░ The job identifier is 300. Jan 11 11:24:48 localhost systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. ░░ Subject: A start job for unit dracut-shutdown.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dracut-shutdown.service has finished successfully. ░░ ░░ The job identifier is 142. Jan 11 11:24:48 localhost (qbalance)[657]: irqbalance.service: Referenced but unset environment variable evaluates to an empty string: IRQBALANCE_ARGS Jan 11 11:24:48 localhost systemd-logind[659]: New seat seat0. ░░ Subject: A new seat seat0 is now available ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new seat seat0 has been configured and is now available. Jan 11 11:24:48 localhost systemd-logind[659]: Watching system buttons on /dev/input/event0 (Power Button) Jan 11 11:24:48 localhost systemd-logind[659]: Watching system buttons on /dev/input/event1 (Sleep Button) Jan 11 11:24:48 localhost systemd-logind[659]: Watching system buttons on /dev/input/event2 (AT Translated Set 2 keyboard) Jan 11 11:24:48 localhost systemd[1]: Started systemd-logind.service - User Login Management. ░░ Subject: A start job for unit systemd-logind.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-logind.service has finished successfully. ░░ ░░ The job identifier is 262. Jan 11 11:24:49 localhost rngd[658]: Disabling 7: PKCS11 Entropy generator (pkcs11) Jan 11 11:24:49 localhost rngd[658]: Disabling 5: NIST Network Entropy Beacon (nist) Jan 11 11:24:49 localhost rngd[658]: Disabling 9: Qrypt quantum entropy beacon (qrypt) Jan 11 11:24:49 localhost rngd[658]: Disabling 10: Named pipe entropy input (namedpipe) Jan 11 11:24:49 localhost rngd[658]: Initializing available sources Jan 11 11:24:49 localhost rngd[658]: [hwrng ]: Initialization Failed Jan 11 11:24:49 localhost rngd[658]: [rdrand]: Enabling RDRAND rng support Jan 11 11:24:49 localhost rngd[658]: [rdrand]: Initialized Jan 11 11:24:49 localhost rngd[658]: [jitter]: JITTER timeout set to 5 sec Jan 11 11:24:49 localhost rngd[658]: [jitter]: Initializing AES buffer Jan 11 11:24:49 localhost chronyd[674]: chronyd version 4.6.1 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Jan 11 11:24:49 localhost chronyd[674]: Frequency 0.000 +/- 1000000.000 ppm read from /var/lib/chrony/drift Jan 11 11:24:49 localhost chronyd[674]: Loaded seccomp filter (level 2) Jan 11 11:24:49 localhost systemd[1]: Started chronyd.service - NTP client/server. ░░ Subject: A start job for unit chronyd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has finished successfully. ░░ ░░ The job identifier is 265. Jan 11 11:24:51 localhost cloud-init[679]: Cloud-init v. 24.1.4-21.el10 running 'init-local' at Sat, 11 Jan 2025 16:24:51 +0000. Up 12.45 seconds. Jan 11 11:24:51 localhost dhcpcd[681]: dhcpcd-10.0.6 starting Jan 11 11:24:51 localhost kernel: 8021q: 802.1Q VLAN Support v1.8 Jan 11 11:24:51 localhost systemd[1]: Listening on systemd-rfkill.socket - Load/Save RF Kill Switch Status /dev/rfkill Watch. ░░ Subject: A start job for unit systemd-rfkill.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-rfkill.socket has finished successfully. ░░ ░░ The job identifier is 309. Jan 11 11:24:51 localhost kernel: cfg80211: Loading compiled-in X.509 certificates for regulatory database Jan 11 11:24:51 localhost kernel: Loaded X.509 cert 'sforshee: 00b28ddf47aef9cea7' Jan 11 11:24:51 localhost kernel: Loaded X.509 cert 'wens: 61c038651aabdcf94bd0ac7ff06c7248db18c600' Jan 11 11:24:51 localhost kernel: platform regulatory.0: Direct firmware load for regulatory.db failed with error -2 Jan 11 11:24:51 localhost kernel: cfg80211: failed to load regulatory.db Jan 11 11:24:51 localhost dhcpcd[684]: DUID 00:01:00:01:2f:15:57:d3:0e:07:a2:74:b2:e7 Jan 11 11:24:51 localhost dhcpcd[684]: eth0: IAID a2:74:b2:e7 Jan 11 11:24:52 localhost dhcpcd[684]: eth0: soliciting a DHCP lease Jan 11 11:24:52 localhost dhcpcd[684]: eth0: offered 10.31.42.117 from 10.31.40.1 Jan 11 11:24:52 localhost dhcpcd[684]: eth0: leased 10.31.42.117 for 3600 seconds Jan 11 11:24:52 localhost dhcpcd[684]: eth0: adding route to 10.31.40.0/22 Jan 11 11:24:52 localhost dhcpcd[684]: eth0: adding default route via 10.31.40.1 Jan 11 11:24:52 localhost dhcpcd[684]: control command: /usr/sbin/dhcpcd --dumplease --ipv4only eth0 Jan 11 11:24:52 localhost systemd[1]: Starting systemd-hostnamed.service - Hostname Service... ░░ Subject: A start job for unit systemd-hostnamed.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has begun execution. ░░ ░░ The job identifier is 318. Jan 11 11:24:52 localhost systemd[1]: Started systemd-hostnamed.service - Hostname Service. ░░ Subject: A start job for unit systemd-hostnamed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has finished successfully. ░░ ░░ The job identifier is 318. Jan 11 11:24:52 ip-10-31-42-117.us-east-1.aws.redhat.com systemd-hostnamed[701]: Hostname set to (static) Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Finished cloud-init-local.service - Initial cloud-init job (pre-networking). ░░ Subject: A start job for unit cloud-init-local.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init-local.service has finished successfully. ░░ ░░ The job identifier is 270. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Reached target network-pre.target - Preparation for Network. ░░ Subject: A start job for unit network-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network-pre.target has finished successfully. ░░ ░░ The job identifier is 197. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting NetworkManager.service - Network Manager... ░░ Subject: A start job for unit NetworkManager.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager.service has begun execution. ░░ ░░ The job identifier is 206. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.2805] NetworkManager (version 1.51.5-1.el10) is starting... (boot:32ea4921-53e5-4ff5-97c7-376caea73887) Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.2807] Read config: /etc/NetworkManager/NetworkManager.conf, /etc/NetworkManager/conf.d/30-cloud-init-ip6-addr-gen-mode.conf Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3003] manager[0x5642fd162d10]: monitoring kernel firmware directory '/lib/firmware'. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3048] hostname: hostname: using hostnamed Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3048] hostname: static hostname changed from (none) to "ip-10-31-42-117.us-east-1.aws.redhat.com" Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3055] dns-mgr: init: dns=default,systemd-resolved rc-manager=symlink (auto) Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3057] manager[0x5642fd162d10]: rfkill: Wi-Fi hardware radio set enabled Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3058] manager[0x5642fd162d10]: rfkill: WWAN hardware radio set enabled Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3120] manager: rfkill: Wi-Fi enabled by radio killswitch; enabled by state file Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3122] manager: rfkill: WWAN enabled by radio killswitch; enabled by state file Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3123] manager: Networking is enabled by state file Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3147] settings: Loaded settings plugin: keyfile (internal) Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3255] dhcp: init: Using DHCP client 'internal' Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3261] manager: (lo): new Loopback device (/org/freedesktop/NetworkManager/Devices/1) Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 396. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3309] device (lo): state change: unmanaged -> unavailable (reason 'connection-assumed', managed-type: 'external') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3347] device (lo): state change: unavailable -> disconnected (reason 'connection-assumed', managed-type: 'external') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3355] device (lo): Activation: starting connection 'lo' (825a6115-524f-48df-bada-def2354bc13b) Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3362] manager: (eth0): new Ethernet device (/org/freedesktop/NetworkManager/Devices/2) Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3385] device (eth0): state change: unmanaged -> unavailable (reason 'managed', managed-type: 'external') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3426] bus-manager: acquired D-Bus service "org.freedesktop.NetworkManager" Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started NetworkManager.service - Network Manager. ░░ Subject: A start job for unit NetworkManager.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager.service has finished successfully. ░░ ░░ The job identifier is 206. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Reached target network.target - Network. ░░ Subject: A start job for unit network.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network.target has finished successfully. ░░ ░░ The job identifier is 209. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3465] device (lo): state change: disconnected -> prepare (reason 'none', managed-type: 'external') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3467] device (lo): state change: prepare -> config (reason 'none', managed-type: 'external') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3469] device (lo): state change: config -> ip-config (reason 'none', managed-type: 'external') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3470] device (eth0): carrier: link connected Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3472] device (lo): state change: ip-config -> ip-check (reason 'none', managed-type: 'external') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3527] device (eth0): state change: unavailable -> disconnected (reason 'carrier-changed', managed-type: 'full') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting NetworkManager-wait-online.service - Network Manager Wait Online... ░░ Subject: A start job for unit NetworkManager-wait-online.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-wait-online.service has begun execution. ░░ ░░ The job identifier is 205. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3568] policy: auto-activating connection 'cloud-init eth0' (1dd9a779-d327-56e1-8454-c65e2556c12c) Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3573] device (eth0): Activation: starting connection 'cloud-init eth0' (1dd9a779-d327-56e1-8454-c65e2556c12c) Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3574] device (eth0): state change: disconnected -> prepare (reason 'none', managed-type: 'full') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3590] manager: NetworkManager state is now CONNECTING Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3591] device (eth0): state change: prepare -> config (reason 'none', managed-type: 'full') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3596] device (eth0): state change: config -> ip-config (reason 'none', managed-type: 'full') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3599] dhcp4 (eth0): activation: beginning transaction (timeout in 45 seconds) Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.3634] dhcp4 (eth0): state changed new lease, address=10.31.42.117, acd pending Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting gssproxy.service - GSSAPI Proxy Daemon... ░░ Subject: A start job for unit gssproxy.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit gssproxy.service has begun execution. ░░ ░░ The job identifier is 244. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 396. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.4502] device (lo): state change: ip-check -> secondaries (reason 'none', managed-type: 'external') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.4513] device (lo): state change: secondaries -> activated (reason 'none', managed-type: 'external') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.4518] device (lo): Activation: successful, device activated. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.5157] dhcp4 (eth0): state changed new lease, address=10.31.42.117 Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.5169] policy: set 'cloud-init eth0' (eth0) as default for IPv4 routing and DNS Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.5215] device (eth0): state change: ip-config -> ip-check (reason 'none', managed-type: 'full') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.5248] device (eth0): state change: ip-check -> secondaries (reason 'none', managed-type: 'full') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.5258] device (eth0): state change: secondaries -> activated (reason 'none', managed-type: 'full') Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.5271] manager: NetworkManager state is now CONNECTED_SITE Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.5277] device (eth0): Activation: successful, device activated. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.5290] manager: NetworkManager state is now CONNECTED_GLOBAL Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com NetworkManager[708]: [1736612693.5310] manager: startup complete Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Finished NetworkManager-wait-online.service - Network Manager Wait Online. ░░ Subject: A start job for unit NetworkManager-wait-online.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-wait-online.service has finished successfully. ░░ ░░ The job identifier is 205. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting cloud-init.service - Initial cloud-init job (metadata service crawler)... ░░ Subject: A start job for unit cloud-init.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.service has begun execution. ░░ ░░ The job identifier is 269. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started gssproxy.service - GSSAPI Proxy Daemon. ░░ Subject: A start job for unit gssproxy.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit gssproxy.service has finished successfully. ░░ ░░ The job identifier is 244. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: rpc-gssd.service - RPC security service for NFS client and server was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). ░░ Subject: A start job for unit rpc-gssd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-gssd.service has finished successfully. ░░ ░░ The job identifier is 245. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Reached target nfs-client.target - NFS client services. ░░ Subject: A start job for unit nfs-client.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit nfs-client.target has finished successfully. ░░ ░░ The job identifier is 242. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. ░░ Subject: A start job for unit remote-fs-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-fs-pre.target has finished successfully. ░░ ░░ The job identifier is 249. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. ░░ Subject: A start job for unit remote-cryptsetup.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-cryptsetup.target has finished successfully. ░░ ░░ The job identifier is 251. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Reached target remote-fs.target - Remote File Systems. ░░ Subject: A start job for unit remote-fs.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-fs.target has finished successfully. ░░ ░░ The job identifier is 275. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: systemd-pcrphase.service - TPM PCR Barrier (User) was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ░░ Subject: A start job for unit systemd-pcrphase.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-pcrphase.service has finished successfully. ░░ ░░ The job identifier is 191. Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com chronyd[674]: Added source 10.11.160.238 Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com chronyd[674]: Added source 10.18.100.10 Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com chronyd[674]: Added source 10.2.32.37 Jan 11 11:24:53 ip-10-31-42-117.us-east-1.aws.redhat.com chronyd[674]: Added source 10.2.32.38 Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com rngd[658]: [jitter]: Unable to obtain AES key, disabling JITTER source Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com rngd[658]: [jitter]: Initialization Failed Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com rngd[658]: Process privileges have been dropped to 2:2 Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: Cloud-init v. 24.1.4-21.el10 running 'init' at Sat, 11 Jan 2025 16:24:54 +0000. Up 15.24 seconds. Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: ++++++++++++++++++++++++++++++++++++++Net device info+++++++++++++++++++++++++++++++++++++++ Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: +--------+------+-----------------------------+---------------+--------+-------------------+ Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: | Device | Up | Address | Mask | Scope | Hw-Address | Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: +--------+------+-----------------------------+---------------+--------+-------------------+ Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: | eth0 | True | 10.31.42.117 | 255.255.252.0 | global | 0e:07:a2:74:b2:e7 | Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: | eth0 | True | fe80::c07:a2ff:fe74:b2e7/64 | . | link | 0e:07:a2:74:b2:e7 | Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: | lo | True | 127.0.0.1 | 255.0.0.0 | host | . | Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: | lo | True | ::1/128 | . | host | . | Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: +--------+------+-----------------------------+---------------+--------+-------------------+ Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: ++++++++++++++++++++++++++++Route IPv4 info+++++++++++++++++++++++++++++ Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: +-------+-------------+------------+---------------+-----------+-------+ Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: | Route | Destination | Gateway | Genmask | Interface | Flags | Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: +-------+-------------+------------+---------------+-----------+-------+ Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: | 0 | 0.0.0.0 | 10.31.40.1 | 0.0.0.0 | eth0 | UG | Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: | 1 | 10.31.40.0 | 0.0.0.0 | 255.255.252.0 | eth0 | U | Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: +-------+-------------+------------+---------------+-----------+-------+ Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: +++++++++++++++++++Route IPv6 info+++++++++++++++++++ Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: +-------+-------------+---------+-----------+-------+ Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: | Route | Destination | Gateway | Interface | Flags | Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: +-------+-------------+---------+-----------+-------+ Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: | 0 | fe80::/64 | :: | eth0 | U | Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: | 2 | multicast | :: | eth0 | U | Jan 11 11:24:54 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: ci-info: +-------+-------------+---------+-----------+-------+ Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: Generating public/private rsa key pair. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: Your identification has been saved in /etc/ssh/ssh_host_rsa_key Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: Your public key has been saved in /etc/ssh/ssh_host_rsa_key.pub Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: The key fingerprint is: Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: SHA256:qZohLExZJgnYodHZriY4A2iKWYA2rXTv66nHBhWizIo root@ip-10-31-42-117.us-east-1.aws.redhat.com Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: The key's randomart image is: Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: +---[RSA 3072]----+ Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: |=oo+ | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: |=*=+.. | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: |B+=+o . | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: |o** .o . | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: |O= .o S | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: |E.o. . . | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: |.*o + o | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | . . O. | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | .Bo | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: +----[SHA256]-----+ Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: Generating public/private ecdsa key pair. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: Your identification has been saved in /etc/ssh/ssh_host_ecdsa_key Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: Your public key has been saved in /etc/ssh/ssh_host_ecdsa_key.pub Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: The key fingerprint is: Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: SHA256:w5oYWT2WwpJa5o/s2thdosdkvPHlxOZDevlb2GzZPKc root@ip-10-31-42-117.us-east-1.aws.redhat.com Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: The key's randomart image is: Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: +---[ECDSA 256]---+ Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | o . . | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | = + = | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | = + + . | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | . +. S. | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | . ==o .* + + | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | ++=+.O .. *.+| Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | = o+oo * o .o| Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | o.=.. . oo.E | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: +----[SHA256]-----+ Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: Generating public/private ed25519 key pair. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: Your identification has been saved in /etc/ssh/ssh_host_ed25519_key Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: Your public key has been saved in /etc/ssh/ssh_host_ed25519_key.pub Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: The key fingerprint is: Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: SHA256:SIvYId5ywlYstBn54+O2ixMuXG2yYEuslpGHxtLY+HI root@ip-10-31-42-117.us-east-1.aws.redhat.com Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: The key's randomart image is: Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: +--[ED25519 256]--+ Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | o. | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | ..= | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | =.+ . | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: | o Bo+ o | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: |o=X.*.o S | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: |+%+Boo | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: |BoBo+. | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: |+=Eoo | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: |.+.ooo | Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[795]: +----[SHA256]-----+ Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Finished cloud-init.service - Initial cloud-init job (metadata service crawler). ░░ Subject: A start job for unit cloud-init.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.service has finished successfully. ░░ ░░ The job identifier is 269. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Reached target cloud-config.target - Cloud-config availability. ░░ Subject: A start job for unit cloud-config.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.target has finished successfully. ░░ ░░ The job identifier is 273. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Reached target network-online.target - Network is Online. ░░ Subject: A start job for unit network-online.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network-online.target has finished successfully. ░░ ░░ The job identifier is 204. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting cloud-config.service - Apply the settings specified in cloud-config... ░░ Subject: A start job for unit cloud-config.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.service has begun execution. ░░ ░░ The job identifier is 272. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting kdump.service - Crash recovery kernel arming... ░░ Subject: A start job for unit kdump.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit kdump.service has begun execution. ░░ ░░ The job identifier is 260. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting restraintd.service - The restraint harness.... ░░ Subject: A start job for unit restraintd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit restraintd.service has begun execution. ░░ ░░ The job identifier is 274. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting rpc-statd-notify.service - Notify NFS peers of a restart... ░░ Subject: A start job for unit rpc-statd-notify.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-statd-notify.service has begun execution. ░░ ░░ The job identifier is 250. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com sm-notify[879]: Version 2.8.2 starting Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting rsyslog.service - System Logging Service... ░░ Subject: A start job for unit rsyslog.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rsyslog.service has begun execution. ░░ ░░ The job identifier is 276. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting sshd.service - OpenSSH server daemon... ░░ Subject: A start job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 253. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com (sshd)[887]: sshd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started restraintd.service - The restraint harness.. ░░ Subject: A start job for unit restraintd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit restraintd.service has finished successfully. ░░ ░░ The job identifier is 274. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started rpc-statd-notify.service - Notify NFS peers of a restart. ░░ Subject: A start job for unit rpc-statd-notify.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-statd-notify.service has finished successfully. ░░ ░░ The job identifier is 250. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com sshd[887]: Server listening on 0.0.0.0 port 22. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com sshd[887]: Server listening on :: port 22. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started sshd.service - OpenSSH server daemon. ░░ Subject: A start job for unit sshd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has finished successfully. ░░ ░░ The job identifier is 253. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started rsyslog.service - System Logging Service. ░░ Subject: A start job for unit rsyslog.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rsyslog.service has finished successfully. ░░ ░░ The job identifier is 276. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com rsyslogd[886]: [origin software="rsyslogd" swVersion="8.2412.0-1.el10" x-pid="886" x-info="https://www.rsyslog.com"] start Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com rsyslogd[886]: imjournal: journal files changed, reloading... [v8.2412.0-1.el10 try https://www.rsyslog.com/e/0 ] Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com restraintd[890]: Listening on http://localhost:8081 Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[927]: Cloud-init v. 24.1.4-21.el10 running 'modules:config' at Sat, 11 Jan 2025 16:24:55 +0000. Up 16.90 seconds. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Stopping sshd.service - OpenSSH server daemon... ░░ Subject: A stop job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 487. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com sshd[887]: Received signal 15; terminating. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: sshd.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit sshd.service has successfully entered the 'dead' state. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Stopped sshd.service - OpenSSH server daemon. ░░ Subject: A stop job for unit sshd.service has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd.service has finished. ░░ ░░ The job identifier is 487 and the job result is done. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Stopped target sshd-keygen.target. ░░ Subject: A stop job for unit sshd-keygen.target has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd-keygen.target has finished. ░░ ░░ The job identifier is 571 and the job result is done. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Stopping sshd-keygen.target... ░░ Subject: A stop job for unit sshd-keygen.target has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd-keygen.target has begun execution. ░░ ░░ The job identifier is 571. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: ssh-host-keys-migration.service - Update OpenSSH host key permissions was skipped because of an unmet condition check (ConditionPathExists=!/var/lib/.ssh-host-keys-migration). ░░ Subject: A start job for unit ssh-host-keys-migration.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit ssh-host-keys-migration.service has finished successfully. ░░ ░░ The job identifier is 565. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: sshd-keygen@ecdsa.service - OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ecdsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ecdsa.service has finished successfully. ░░ ░░ The job identifier is 567. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: sshd-keygen@ed25519.service - OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ed25519.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ed25519.service has finished successfully. ░░ ░░ The job identifier is 569. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: sshd-keygen@rsa.service - OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@rsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@rsa.service has finished successfully. ░░ ░░ The job identifier is 570. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Reached target sshd-keygen.target. ░░ Subject: A start job for unit sshd-keygen.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen.target has finished successfully. ░░ ░░ The job identifier is 571. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting sshd.service - OpenSSH server daemon... ░░ Subject: A start job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 487. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com (sshd)[932]: sshd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com sshd[932]: Server listening on 0.0.0.0 port 22. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com sshd[932]: Server listening on :: port 22. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started sshd.service - OpenSSH server daemon. ░░ Subject: A start job for unit sshd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has finished successfully. ░░ ░░ The job identifier is 487. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Finished cloud-config.service - Apply the settings specified in cloud-config. ░░ Subject: A start job for unit cloud-config.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.service has finished successfully. ░░ ░░ The job identifier is 272. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting cloud-final.service - Execute cloud user/final scripts... ░░ Subject: A start job for unit cloud-final.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-final.service has begun execution. ░░ ░░ The job identifier is 271. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... ░░ Subject: A start job for unit systemd-user-sessions.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-user-sessions.service has begun execution. ░░ ░░ The job identifier is 252. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. ░░ Subject: A start job for unit systemd-user-sessions.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-user-sessions.service has finished successfully. ░░ ░░ The job identifier is 252. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started crond.service - Command Scheduler. ░░ Subject: A start job for unit crond.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit crond.service has finished successfully. ░░ ░░ The job identifier is 277. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started getty@tty1.service - Getty on tty1. ░░ Subject: A start job for unit getty@tty1.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit getty@tty1.service has finished successfully. ░░ ░░ The job identifier is 234. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. ░░ Subject: A start job for unit serial-getty@ttyS0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit serial-getty@ttyS0.service has finished successfully. ░░ ░░ The job identifier is 229. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Reached target getty.target - Login Prompts. ░░ Subject: A start job for unit getty.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit getty.target has finished successfully. ░░ ░░ The job identifier is 228. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Reached target multi-user.target - Multi-User System. ░░ Subject: A start job for unit multi-user.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit multi-user.target has finished successfully. ░░ ░░ The job identifier is 121. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting systemd-update-utmp-runlevel.service - Record Runlevel Change in UTMP... ░░ Subject: A start job for unit systemd-update-utmp-runlevel.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp-runlevel.service has begun execution. ░░ ░░ The job identifier is 240. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com crond[990]: (CRON) STARTUP (1.7.0) Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com crond[990]: (CRON) INFO (Syslog will be used instead of sendmail.) Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com crond[990]: (CRON) INFO (RANDOM_DELAY will be scaled with factor 80% if used.) Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com crond[990]: (CRON) INFO (running with inotify support) Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-update-utmp-runlevel.service has successfully entered the 'dead' state. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Finished systemd-update-utmp-runlevel.service - Record Runlevel Change in UTMP. ░░ Subject: A start job for unit systemd-update-utmp-runlevel.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp-runlevel.service has finished successfully. ░░ ░░ The job identifier is 240. Jan 11 11:24:55 ip-10-31-42-117.us-east-1.aws.redhat.com kdumpctl[883]: kdump: Detected change(s) in the following file(s): /etc/fstab Jan 11 11:24:56 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[1105]: Cloud-init v. 24.1.4-21.el10 running 'modules:final' at Sat, 11 Jan 2025 16:24:56 +0000. Up 17.41 seconds. Jan 11 11:24:56 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[1145]: ############################################################# Jan 11 11:24:56 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[1146]: -----BEGIN SSH HOST KEY FINGERPRINTS----- Jan 11 11:24:56 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[1148]: 256 SHA256:w5oYWT2WwpJa5o/s2thdosdkvPHlxOZDevlb2GzZPKc root@ip-10-31-42-117.us-east-1.aws.redhat.com (ECDSA) Jan 11 11:24:56 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[1150]: 256 SHA256:SIvYId5ywlYstBn54+O2ixMuXG2yYEuslpGHxtLY+HI root@ip-10-31-42-117.us-east-1.aws.redhat.com (ED25519) Jan 11 11:24:56 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[1152]: 3072 SHA256:qZohLExZJgnYodHZriY4A2iKWYA2rXTv66nHBhWizIo root@ip-10-31-42-117.us-east-1.aws.redhat.com (RSA) Jan 11 11:24:56 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[1153]: -----END SSH HOST KEY FINGERPRINTS----- Jan 11 11:24:56 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[1154]: ############################################################# Jan 11 11:24:56 ip-10-31-42-117.us-east-1.aws.redhat.com cloud-init[1105]: Cloud-init v. 24.1.4-21.el10 finished at Sat, 11 Jan 2025 16:24:56 +0000. Datasource DataSourceEc2Local. Up 17.55 seconds Jan 11 11:24:56 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Finished cloud-final.service - Execute cloud user/final scripts. ░░ Subject: A start job for unit cloud-final.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-final.service has finished successfully. ░░ ░░ The job identifier is 271. Jan 11 11:24:56 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Reached target cloud-init.target - Cloud-init target. ░░ Subject: A start job for unit cloud-init.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.target has finished successfully. ░░ ░░ The job identifier is 268. Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com kernel: block xvda: the capability attribute has been deprecated. Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com kdumpctl[883]: kdump: Rebuilding /boot/initramfs-6.12.0-38.el10.x86_64kdump.img Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1395]: dracut-103-1.el10 Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 0 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 0 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 48 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 48 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 49 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 49 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 50 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 50 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 51 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 51 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 52 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 52 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 53 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 53 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 54 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 54 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 55 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 55 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 56 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 56 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 57 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 57 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 58 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 58 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: Cannot change IRQ 59 affinity: Permission denied Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com irqbalance[657]: IRQ 59 affinity is now unmanaged Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1398]: Executing: /usr/bin/dracut --list-modules Jan 11 11:24:58 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1469]: dracut-103-1.el10 Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Executing: /usr/bin/dracut --add kdumpbase --quiet --hostonly --hostonly-cmdline --hostonly-i18n --hostonly-mode strict --hostonly-nics --aggressive-strip --omit "rdma plymouth resume ifcfg earlykdump" --mount "/dev/disk/by-uuid/5421f911-fafd-4f0d-bf2e-2916252992eb /sysroot xfs rw,relatime,seclabel,attr2,inode64,logbufs=8,logbsize=32k,noquota" --add squash-squashfs --squash-compressor zstd --no-hostonly-default-device -f /boot/initramfs-6.12.0-38.el10.x86_64kdump.img 6.12.0-38.el10.x86_64 Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-bsod' will not be installed, because command '/usr/lib/systemd/systemd-bsod' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-networkd' will not be installed, because command '/usr/lib/systemd/systemd-networkd' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-networkd' will not be installed, because command '/usr/lib/systemd/systemd-networkd-wait-online' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-pcrphase' will not be installed, because command '/usr/lib/systemd/systemd-pcrphase' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-portabled' will not be installed, because command 'portablectl' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-portabled' will not be installed, because command '/usr/lib/systemd/systemd-portabled' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-resolved' will not be installed, because command '/usr/lib/systemd/systemd-resolved' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-timesyncd' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-time-wait-sync' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'busybox' will not be installed, because command 'busybox' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'dbus-daemon' will not be installed, because command 'dbus-daemon' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'connman' will not be installed, because command 'connmand' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'connman' will not be installed, because command 'connmanctl' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'connman' will not be installed, because command 'connmand-wait-online' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'ifcfg' will not be installed, because it's in the list to be omitted! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'plymouth' will not be installed, because it's in the list to be omitted! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: 62bluetooth: Could not find any command of '/usr/lib/bluetooth/bluetoothd /usr/libexec/bluetooth/bluetoothd'! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'btrfs' will not be installed, because command 'btrfs' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'dmraid' will not be installed, because command 'dmraid' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'mdraid' will not be installed, because command 'mdadm' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'multipath' will not be installed, because command 'multipath' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'crypt-gpg' will not be installed, because command 'gpg' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'pcsc' will not be installed, because command 'pcscd' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'iscsi' will not be installed, because command 'iscsid' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'nvmf' will not be installed, because command 'nvme' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'resume' will not be installed, because it's in the list to be omitted! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'squash-erofs' will not be installed, because command 'mkfs.erofs' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'squash-erofs' will not be installed, because command 'fsck.erofs' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'biosdevname' will not be installed, because command 'biosdevname' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'earlykdump' will not be installed, because it's in the list to be omitted! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-bsod' will not be installed, because command '/usr/lib/systemd/systemd-bsod' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-pcrphase' will not be installed, because command '/usr/lib/systemd/systemd-pcrphase' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-portabled' will not be installed, because command 'portablectl' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-portabled' will not be installed, because command '/usr/lib/systemd/systemd-portabled' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-resolved' will not be installed, because command '/usr/lib/systemd/systemd-resolved' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-timesyncd' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-time-wait-sync' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'busybox' will not be installed, because command 'busybox' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'dbus-daemon' will not be installed, because command 'dbus-daemon' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'connman' will not be installed, because command 'connmand' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'connman' will not be installed, because command 'connmanctl' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'connman' will not be installed, because command 'connmand-wait-online' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: 62bluetooth: Could not find any command of '/usr/lib/bluetooth/bluetoothd /usr/libexec/bluetooth/bluetoothd'! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'btrfs' will not be installed, because command 'btrfs' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'dmraid' will not be installed, because command 'dmraid' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'mdraid' will not be installed, because command 'mdadm' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'multipath' will not be installed, because command 'multipath' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'crypt-gpg' will not be installed, because command 'gpg' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'pcsc' will not be installed, because command 'pcscd' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'iscsi' will not be installed, because command 'iscsid' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'nvmf' will not be installed, because command 'nvme' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'squash-erofs' will not be installed, because command 'mkfs.erofs' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Module 'squash-erofs' will not be installed, because command 'fsck.erofs' could not be found! Jan 11 11:24:59 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: systemd *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: fips *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com chronyd[674]: Selected source 10.2.32.38 Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: fips-crypto-policies *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: systemd-ask-password *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: systemd-initrd *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: systemd-journald *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: systemd-modules-load *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: systemd-sysctl *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: systemd-sysusers *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: systemd-tmpfiles *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: systemd-udevd *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: rngd *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: i18n *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: drm *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: prefixdevname *** Jan 11 11:25:00 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: kernel-modules *** Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: kernel-modules-extra *** Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: kernel-modules-extra: configuration source "/run/depmod.d" does not exist Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: kernel-modules-extra: configuration source "/lib/depmod.d" does not exist Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: kernel-modules-extra: parsing configuration file "/etc/depmod.d/dist.conf" Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: kernel-modules-extra: /etc/depmod.d/dist.conf: added "updates extra built-in weak-updates" to the list of search directories Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: pcmcia *** Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Skipping udev rule: 60-pcmcia.rules Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: fstab-sys *** Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: hwdb *** Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: rootfs-block *** Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: squash-squashfs *** Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: terminfo *** Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: udev-rules *** Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: dracut-systemd *** Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: usrmount *** Jan 11 11:25:01 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: base *** Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: fs-lib *** Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: kdumpbase *** Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: memstrack *** Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: microcode_ctl-fw_dir_override *** Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: microcode_ctl module: mangling fw_dir Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: microcode_ctl: reset fw_dir to "/lib/firmware/updates /lib/firmware" Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel"... Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: microcode_ctl: intel: caveats check for kernel version "6.12.0-38.el10.x86_64" passed, adding "/usr/share/microcode_ctl/ucode_with_caveats/intel" to fw_dir variable Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-4f-01"... Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: microcode_ctl: configuration "intel-06-4f-01" is ignored Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: microcode_ctl: final fw_dir: "/usr/share/microcode_ctl/ucode_with_caveats/intel /lib/firmware/updates /lib/firmware" Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: shutdown *** Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including module: squash-lib *** Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Including modules done *** Jan 11 11:25:02 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Installing kernel module dependencies *** Jan 11 11:25:03 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Installing kernel module dependencies done *** Jan 11 11:25:03 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Resolving executable dependencies *** Jan 11 11:25:03 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Jan 11 11:25:03 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Resolving executable dependencies done *** Jan 11 11:25:03 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Hardlinking files *** Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Mode: real Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Method: sha256 Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Files: 548 Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Linked: 23 files Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Compared: 0 xattrs Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Compared: 51 files Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Saved: 13.58 MiB Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Duration: 0.193814 seconds Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Hardlinking files done *** Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Generating early-microcode cpio image *** Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Constructing GenuineIntel.bin *** Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Constructing GenuineIntel.bin *** Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Store current command line parameters *** Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: Stored kernel commandline: Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: No dracut internal kernel commandline stored in the initramfs Jan 11 11:25:04 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Squashing the files inside the initramfs *** Jan 11 11:25:12 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Squashing the files inside the initramfs done *** Jan 11 11:25:12 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Creating image file '/boot/initramfs-6.12.0-38.el10.x86_64kdump.img' *** Jan 11 11:25:12 ip-10-31-42-117.us-east-1.aws.redhat.com dracut[1472]: *** Creating initramfs image file '/boot/initramfs-6.12.0-38.el10.x86_64kdump.img' done *** Jan 11 11:25:12 ip-10-31-42-117.us-east-1.aws.redhat.com kernel: PKCS7: Message signed outside of X.509 validity window Jan 11 11:25:13 ip-10-31-42-117.us-east-1.aws.redhat.com kdumpctl[883]: kdump: kexec: loaded kdump kernel Jan 11 11:25:13 ip-10-31-42-117.us-east-1.aws.redhat.com kdumpctl[883]: kdump: Starting kdump: [OK] Jan 11 11:25:13 ip-10-31-42-117.us-east-1.aws.redhat.com kdumpctl[883]: kdump: Notice: No vmcore creation test performed! Jan 11 11:25:13 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Finished kdump.service - Crash recovery kernel arming. ░░ Subject: A start job for unit kdump.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit kdump.service has finished successfully. ░░ ░░ The job identifier is 260. Jan 11 11:25:13 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Startup finished in 986ms (kernel) + 4.099s (initrd) + 29.406s (userspace) = 34.492s. ░░ Subject: System start-up is now complete ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ All system services necessary queued for starting at boot have been ░░ started. Note that this does not mean that the machine is now idle as services ░░ might still be busy with completing start-up. ░░ ░░ Kernel start-up required 986465 microseconds. ░░ ░░ Initrd start-up required 4099369 microseconds. ░░ ░░ Userspace start-up required 29406330 microseconds. Jan 11 11:25:23 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: systemd-hostnamed.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-hostnamed.service has successfully entered the 'dead' state. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com sshd-session[4334]: Accepted publickey for root from 10.30.34.106 port 36526 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd-logind[659]: New session 1 of user root. ░░ Subject: A new session 1 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 1 has been created for the user root. ░░ ░░ The leading process of the session is 4334. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Created slice user-0.slice - User Slice of UID 0. ░░ Subject: A start job for unit user-0.slice has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-0.slice has finished successfully. ░░ ░░ The job identifier is 650. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting user-runtime-dir@0.service - User Runtime Directory /run/user/0... ░░ Subject: A start job for unit user-runtime-dir@0.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-runtime-dir@0.service has begun execution. ░░ ░░ The job identifier is 572. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Finished user-runtime-dir@0.service - User Runtime Directory /run/user/0. ░░ Subject: A start job for unit user-runtime-dir@0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-runtime-dir@0.service has finished successfully. ░░ ░░ The job identifier is 572. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting user@0.service - User Manager for UID 0... ░░ Subject: A start job for unit user@0.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user@0.service has begun execution. ░░ ░░ The job identifier is 652. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd-logind[659]: New session 2 of user root. ░░ Subject: A new session 2 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 2 has been created for the user root. ░░ ░░ The leading process of the session is 4339. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com (systemd)[4339]: pam_unix(systemd-user:session): session opened for user root(uid=0) by root(uid=0) Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Queued start job for default target default.target. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Created slice app.slice - User Application Slice. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 4. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: grub-boot-success.timer - Mark boot as successful after the user session has run 2 minutes was skipped because of an unmet condition check (ConditionUser=!@system). ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 9. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of User's Temporary Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 10. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Reached target paths.target - Paths. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 7. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Reached target timers.target - Timers. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 8. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Starting dbus.socket - D-Bus User Message Bus Socket... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 12. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Starting systemd-tmpfiles-setup.service - Create User Files and Directories... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 3. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Finished systemd-tmpfiles-setup.service - Create User Files and Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 3. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Listening on dbus.socket - D-Bus User Message Bus Socket. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 12. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Reached target sockets.target - Sockets. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 11. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Reached target basic.target - Basic System. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 2. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Reached target default.target - Main User Target. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 1. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[4339]: Startup finished in 113ms. ░░ Subject: User manager start-up is now complete ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The user manager instance for user 0 has been started. All services queued ░░ for starting have been started. Note that other services might still be starting ░░ up or be started at any later time. ░░ ░░ Startup of the manager took 113388 microseconds. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started user@0.service - User Manager for UID 0. ░░ Subject: A start job for unit user@0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user@0.service has finished successfully. ░░ ░░ The job identifier is 652. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started session-1.scope - Session 1 of User root. ░░ Subject: A start job for unit session-1.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-1.scope has finished successfully. ░░ ░░ The job identifier is 733. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com sshd-session[4334]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com sshd-session[4350]: Received disconnect from 10.30.34.106 port 36526:11: disconnected by user Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com sshd-session[4350]: Disconnected from user root 10.30.34.106 port 36526 Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com sshd-session[4334]: pam_unix(sshd:session): session closed for user root Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: session-1.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-1.scope has successfully entered the 'dead' state. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd-logind[659]: Session 1 logged out. Waiting for processes to exit. Jan 11 11:27:39 ip-10-31-42-117.us-east-1.aws.redhat.com systemd-logind[659]: Removed session 1. ░░ Subject: Session 1 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 1 has been terminated. Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com sshd-session[4386]: Accepted publickey for root from 10.31.9.220 port 38948 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com sshd-session[4387]: Accepted publickey for root from 10.31.9.220 port 38960 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com systemd-logind[659]: New session 3 of user root. ░░ Subject: A new session 3 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 3 has been created for the user root. ░░ ░░ The leading process of the session is 4386. Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started session-3.scope - Session 3 of User root. ░░ Subject: A start job for unit session-3.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-3.scope has finished successfully. ░░ ░░ The job identifier is 815. Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com systemd-logind[659]: New session 4 of user root. ░░ Subject: A new session 4 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 4 has been created for the user root. ░░ ░░ The leading process of the session is 4387. Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started session-4.scope - Session 4 of User root. ░░ Subject: A start job for unit session-4.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-4.scope has finished successfully. ░░ ░░ The job identifier is 897. Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com sshd-session[4386]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com sshd-session[4387]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com sshd-session[4393]: Received disconnect from 10.31.9.220 port 38960:11: disconnected by user Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com sshd-session[4393]: Disconnected from user root 10.31.9.220 port 38960 Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com sshd-session[4387]: pam_unix(sshd:session): session closed for user root Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: session-4.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-4.scope has successfully entered the 'dead' state. Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com systemd-logind[659]: Session 4 logged out. Waiting for processes to exit. Jan 11 11:27:46 ip-10-31-42-117.us-east-1.aws.redhat.com systemd-logind[659]: Removed session 4. ░░ Subject: Session 4 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 4 has been terminated. Jan 11 11:28:02 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Starting systemd-hostnamed.service - Hostname Service... ░░ Subject: A start job for unit systemd-hostnamed.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has begun execution. ░░ ░░ The job identifier is 979. Jan 11 11:28:02 ip-10-31-42-117.us-east-1.aws.redhat.com systemd[1]: Started systemd-hostnamed.service - Hostname Service. ░░ Subject: A start job for unit systemd-hostnamed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has finished successfully. ░░ ░░ The job identifier is 979. Jan 11 11:28:02 managed-node2 systemd-hostnamed[5841]: Hostname set to (static) Jan 11 11:28:02 managed-node2 NetworkManager[708]: [1736612882.1934] hostname: static hostname changed from "ip-10-31-42-117.us-east-1.aws.redhat.com" to "managed-node2" Jan 11 11:28:02 managed-node2 systemd[1]: Starting NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 1057. Jan 11 11:28:02 managed-node2 systemd[1]: Started NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 1057. Jan 11 11:28:12 managed-node2 systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Jan 11 11:28:32 managed-node2 systemd[1]: systemd-hostnamed.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-hostnamed.service has successfully entered the 'dead' state. Jan 11 11:28:35 managed-node2 sshd-session[6514]: Accepted publickey for root from 10.31.14.128 port 52846 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Jan 11 11:28:35 managed-node2 systemd-logind[659]: New session 5 of user root. ░░ Subject: A new session 5 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 5 has been created for the user root. ░░ ░░ The leading process of the session is 6514. Jan 11 11:28:35 managed-node2 systemd[1]: Started session-5.scope - Session 5 of User root. ░░ Subject: A start job for unit session-5.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-5.scope has finished successfully. ░░ ░░ The job identifier is 1136. Jan 11 11:28:35 managed-node2 sshd-session[6514]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Jan 11 11:28:36 managed-node2 python3.12[6663]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Jan 11 11:28:37 managed-node2 python3.12[6817]: ansible-tempfile Invoked with state=directory prefix=lsr_ suffix=_podman path=None Jan 11 11:28:38 managed-node2 python3.12[6942]: ansible-file Invoked with path=/tmp/lsr_adhv0_3l_podman mode=0777 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False state=None _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:28:39 managed-node2 python3.12[7067]: ansible-user Invoked with name=podman_basic_user uid=3001 state=present non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on managed-node2 update_password=always group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None password_expire_warn=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None Jan 11 11:28:39 managed-node2 useradd[7069]: new group: name=podman_basic_user, GID=3001 Jan 11 11:28:39 managed-node2 useradd[7069]: new user: name=podman_basic_user, UID=3001, GID=3001, home=/home/podman_basic_user, shell=/bin/bash, from=/dev/pts/0 Jan 11 11:28:40 managed-node2 python3.12[7194]: ansible-file Invoked with path=/tmp/lsr_adhv0_3l_podman/httpd1 state=directory mode=0755 owner=podman_basic_user recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:28:40 managed-node2 python3.12[7319]: ansible-file Invoked with path=/tmp/lsr_adhv0_3l_podman/httpd2 state=directory mode=0755 owner=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:28:41 managed-node2 python3.12[7444]: ansible-file Invoked with path=/tmp/lsr_adhv0_3l_podman/httpd3 state=directory mode=0755 owner=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:28:41 managed-node2 python3.12[7569]: ansible-ansible.legacy.stat Invoked with path=/tmp/lsr_adhv0_3l_podman/httpd1/index.txt follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jan 11 11:28:41 managed-node2 python3.12[7669]: ansible-ansible.legacy.copy Invoked with dest=/tmp/lsr_adhv0_3l_podman/httpd1/index.txt mode=0644 owner=podman_basic_user src=/root/.ansible/tmp/ansible-tmp-1736612921.1510544-7094-193496919768865/.source.txt _original_basename=.060zt5t3 follow=False checksum=40bd001563085fc35165329ea1ff5c5ecbdbbeef backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:28:42 managed-node2 python3.12[7794]: ansible-ansible.legacy.stat Invoked with path=/tmp/lsr_adhv0_3l_podman/httpd2/index.txt follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jan 11 11:28:42 managed-node2 python3.12[7894]: ansible-ansible.legacy.copy Invoked with dest=/tmp/lsr_adhv0_3l_podman/httpd2/index.txt mode=0644 owner=root src=/root/.ansible/tmp/ansible-tmp-1736612921.821283-7094-184250787704089/.source.txt _original_basename=.wa9mfvw9 follow=False checksum=40bd001563085fc35165329ea1ff5c5ecbdbbeef backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:28:42 managed-node2 python3.12[8019]: ansible-ansible.legacy.stat Invoked with path=/tmp/lsr_adhv0_3l_podman/httpd3/index.txt follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jan 11 11:28:43 managed-node2 python3.12[8119]: ansible-ansible.legacy.copy Invoked with dest=/tmp/lsr_adhv0_3l_podman/httpd3/index.txt mode=0644 owner=root src=/root/.ansible/tmp/ansible-tmp-1736612922.4798434-7094-30966835373969/.source.txt _original_basename=.8cy2dulw follow=False checksum=40bd001563085fc35165329ea1ff5c5ecbdbbeef backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:28:43 managed-node2 python3.12[8244]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:28:43 managed-node2 python3.12[8369]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:28:45 managed-node2 sudo[8619]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-anvwlwokizgjnjnjtuosodawsgjcavfv ; /usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1736612925.1132412-7210-124838245355389/AnsiballZ_dnf.py' Jan 11 11:28:45 managed-node2 sudo[8619]: pam_unix(sudo:session): session opened for user root(uid=0) by root(uid=0) Jan 11 11:28:45 managed-node2 python3.12[8622]: ansible-ansible.legacy.dnf Invoked with name=['iptables-nft', 'podman', 'shadow-utils-subid'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:28:54 managed-node2 kernel: SELinux: Converting 385 SID table entries... Jan 11 11:28:54 managed-node2 kernel: SELinux: policy capability network_peer_controls=1 Jan 11 11:28:54 managed-node2 kernel: SELinux: policy capability open_perms=1 Jan 11 11:28:54 managed-node2 kernel: SELinux: policy capability extended_socket_class=1 Jan 11 11:28:54 managed-node2 kernel: SELinux: policy capability always_check_network=0 Jan 11 11:28:54 managed-node2 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 11 11:28:54 managed-node2 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 11 11:28:54 managed-node2 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Jan 11 11:28:54 managed-node2 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 11 11:28:54 managed-node2 kernel: SELinux: policy capability userspace_initial_context=0 Jan 11 11:28:57 managed-node2 kernel: SELinux: Converting 386 SID table entries... Jan 11 11:28:57 managed-node2 kernel: SELinux: policy capability network_peer_controls=1 Jan 11 11:28:57 managed-node2 kernel: SELinux: policy capability open_perms=1 Jan 11 11:28:57 managed-node2 kernel: SELinux: policy capability extended_socket_class=1 Jan 11 11:28:57 managed-node2 kernel: SELinux: policy capability always_check_network=0 Jan 11 11:28:57 managed-node2 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 11 11:28:57 managed-node2 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 11 11:28:57 managed-node2 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Jan 11 11:28:57 managed-node2 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 11 11:28:57 managed-node2 kernel: SELinux: policy capability userspace_initial_context=0 Jan 11 11:28:57 managed-node2 setsebool[8707]: The virt_use_nfs policy boolean was changed to 1 by root Jan 11 11:28:57 managed-node2 setsebool[8707]: The virt_sandbox_use_all_caps policy boolean was changed to 1 by root Jan 11 11:29:00 managed-node2 kernel: SELinux: Converting 393 SID table entries... Jan 11 11:29:00 managed-node2 kernel: SELinux: policy capability network_peer_controls=1 Jan 11 11:29:00 managed-node2 kernel: SELinux: policy capability open_perms=1 Jan 11 11:29:00 managed-node2 kernel: SELinux: policy capability extended_socket_class=1 Jan 11 11:29:00 managed-node2 kernel: SELinux: policy capability always_check_network=0 Jan 11 11:29:00 managed-node2 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 11 11:29:00 managed-node2 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 11 11:29:00 managed-node2 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Jan 11 11:29:00 managed-node2 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 11 11:29:00 managed-node2 kernel: SELinux: policy capability userspace_initial_context=0 Jan 11 11:29:01 managed-node2 groupadd[8727]: group added to /etc/group: name=polkitd, GID=114 Jan 11 11:29:01 managed-node2 groupadd[8727]: group added to /etc/gshadow: name=polkitd Jan 11 11:29:01 managed-node2 groupadd[8727]: new group: name=polkitd, GID=114 Jan 11 11:29:01 managed-node2 useradd[8730]: new user: name=polkitd, UID=114, GID=114, home=/, shell=/sbin/nologin, from=none Jan 11 11:29:01 managed-node2 dbus-broker-launch[636]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Jan 11 11:29:01 managed-node2 dbus-broker-launch[636]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Jan 11 11:29:01 managed-node2 systemd[1]: Listening on pcscd.socket - PC/SC Smart Card Daemon Activation Socket. ░░ Subject: A start job for unit pcscd.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit pcscd.socket has finished successfully. ░░ ░░ The job identifier is 1220. Jan 11 11:29:02 managed-node2 systemd[1]: Started run-p9129-i9429.service - [systemd-run] /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-p9129-i9429.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-p9129-i9429.service has finished successfully. ░░ ░░ The job identifier is 1298. Jan 11 11:29:02 managed-node2 systemctl[9130]: Warning: The unit file, source configuration file or drop-ins of man-db-cache-update.service changed on disk. Run 'systemctl daemon-reload' to reload units. Jan 11 11:29:02 managed-node2 systemd[1]: Reload requested from client PID 9133 ('systemctl') (unit session-5.scope)... Jan 11 11:29:02 managed-node2 systemd[1]: Reloading... Jan 11 11:29:02 managed-node2 systemd-ssh-generator[9182]: Failed to query local AF_VSOCK CID: Permission denied Jan 11 11:29:02 managed-node2 systemd-rc-local-generator[9179]: /etc/rc.d/rc.local is not marked executable, skipping. Jan 11 11:29:02 managed-node2 (sd-exec-[9152]: /usr/lib/systemd/system-generators/systemd-ssh-generator failed with exit status 1. Jan 11 11:29:02 managed-node2 systemd[1]: Reloading finished in 195 ms. Jan 11 11:29:02 managed-node2 systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1376. Jan 11 11:29:02 managed-node2 systemd[1]: Queuing reload/restart jobs for marked units… Jan 11 11:29:02 managed-node2 systemd[4339]: Received SIGRTMIN+25 from PID 1 (systemd). Jan 11 11:29:02 managed-node2 systemd[1]: Reloading user@0.service - User Manager for UID 0... ░░ Subject: A reload job for unit user@0.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A reload job for unit user@0.service has begun execution. ░░ ░░ The job identifier is 1454. Jan 11 11:29:02 managed-node2 systemd[4339]: Reexecuting. Jan 11 11:29:02 managed-node2 systemd[1]: Reloaded user@0.service - User Manager for UID 0. ░░ Subject: A reload job for unit user@0.service has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A reload job for unit user@0.service has finished. ░░ ░░ The job identifier is 1454 and the job result is done. Jan 11 11:29:03 managed-node2 sudo[8619]: pam_unix(sudo:session): session closed for user root Jan 11 11:29:04 managed-node2 python3.12[9871]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jan 11 11:29:05 managed-node2 python3.12[10008]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Jan 11 11:29:05 managed-node2 systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Jan 11 11:29:05 managed-node2 systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1376. Jan 11 11:29:05 managed-node2 systemd[1]: run-p9129-i9429.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-p9129-i9429.service has successfully entered the 'dead' state. Jan 11 11:29:05 managed-node2 python3.12[10144]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:29:06 managed-node2 python3.12[10277]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:29:07 managed-node2 python3.12[10408]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:29:07 managed-node2 python3.12[10539]: ansible-ansible.legacy.dnf Invoked with name=['firewalld'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:29:09 managed-node2 python3.12[10671]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False scope=system no_block=False state=None enabled=None force=None Jan 11 11:29:09 managed-node2 python3.12[10804]: ansible-ansible.legacy.systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Jan 11 11:29:09 managed-node2 systemd[1]: Reload requested from client PID 10807 ('systemctl') (unit session-5.scope)... Jan 11 11:29:09 managed-node2 systemd[1]: Reloading... Jan 11 11:29:10 managed-node2 systemd-rc-local-generator[10845]: /etc/rc.d/rc.local is not marked executable, skipping. Jan 11 11:29:10 managed-node2 systemd-ssh-generator[10851]: Failed to query local AF_VSOCK CID: Permission denied Jan 11 11:29:10 managed-node2 (sd-exec-[10825]: /usr/lib/systemd/system-generators/systemd-ssh-generator failed with exit status 1. Jan 11 11:29:10 managed-node2 systemd[1]: Reloading finished in 188 ms. Jan 11 11:29:10 managed-node2 systemd[1]: Starting firewalld.service - firewalld - dynamic firewall daemon... ░░ Subject: A start job for unit firewalld.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit firewalld.service has begun execution. ░░ ░░ The job identifier is 1455. Jan 11 11:29:10 managed-node2 systemd[1]: Started firewalld.service - firewalld - dynamic firewall daemon. ░░ Subject: A start job for unit firewalld.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit firewalld.service has finished successfully. ░░ ░░ The job identifier is 1455. Jan 11 11:29:10 managed-node2 kernel: Warning: Unmaintained driver is detected: ip_set Jan 11 11:29:10 managed-node2 systemd[1]: Starting polkit.service - Authorization Manager... ░░ Subject: A start job for unit polkit.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit polkit.service has begun execution. ░░ ░░ The job identifier is 1538. Jan 11 11:29:10 managed-node2 polkitd[10993]: Started polkitd version 125 Jan 11 11:29:10 managed-node2 systemd[1]: Started polkit.service - Authorization Manager. ░░ Subject: A start job for unit polkit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit polkit.service has finished successfully. ░░ ░░ The job identifier is 1538. Jan 11 11:29:11 managed-node2 python3.12[11045]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['15001-15003/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Jan 11 11:29:12 managed-node2 python3.12[11177]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:29:13 managed-node2 python3.12[11308]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:29:13 managed-node2 python3.12[11439]: ansible-ansible.legacy.dnf Invoked with name=['python3-libselinux', 'python3-policycoreutils'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:29:14 managed-node2 python3.12[11571]: ansible-ansible.legacy.dnf Invoked with name=['policycoreutils-python-utils'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:29:15 managed-node2 dbus-broker-launch[636]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Jan 11 11:29:15 managed-node2 dbus-broker-launch[636]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Jan 11 11:29:15 managed-node2 systemd[1]: Started run-p11577-i11877.service - [systemd-run] /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-p11577-i11877.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-p11577-i11877.service has finished successfully. ░░ ░░ The job identifier is 1619. Jan 11 11:29:15 managed-node2 systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1697. Jan 11 11:29:15 managed-node2 systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Jan 11 11:29:15 managed-node2 systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1697. Jan 11 11:29:15 managed-node2 systemd[1]: run-p11577-i11877.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-p11577-i11877.service has successfully entered the 'dead' state. Jan 11 11:29:15 managed-node2 python3.12[11711]: ansible-setup Invoked with filter=['ansible_selinux'] gather_subset=['all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Jan 11 11:29:17 managed-node2 python3.12[11875]: ansible-fedora.linux_system_roles.local_seport Invoked with ports=['15001-15003'] proto=tcp setype=http_port_t state=present local=False ignore_selinux_state=False reload=True Jan 11 11:29:18 managed-node2 kernel: SELinux: Converting 423 SID table entries... Jan 11 11:29:18 managed-node2 kernel: SELinux: policy capability network_peer_controls=1 Jan 11 11:29:18 managed-node2 kernel: SELinux: policy capability open_perms=1 Jan 11 11:29:18 managed-node2 kernel: SELinux: policy capability extended_socket_class=1 Jan 11 11:29:18 managed-node2 kernel: SELinux: policy capability always_check_network=0 Jan 11 11:29:18 managed-node2 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 11 11:29:18 managed-node2 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 11 11:29:18 managed-node2 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Jan 11 11:29:18 managed-node2 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 11 11:29:18 managed-node2 kernel: SELinux: policy capability userspace_initial_context=0 Jan 11 11:29:18 managed-node2 python3.12[12010]: ansible-fedora.linux_system_roles.selinux_modules_facts Invoked Jan 11 11:29:23 managed-node2 python3.12[12141]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:29:24 managed-node2 python3.12[12274]: ansible-stat Invoked with path=/etc/containers/ansible-kubernetes.d/nopull.yml follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:29:25 managed-node2 python3.12[12405]: ansible-file Invoked with path=/etc/containers/ansible-kubernetes.d state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:29:25 managed-node2 python3.12[12536]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/ansible-kubernetes.d/nopull.yml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jan 11 11:29:25 managed-node2 python3.12[12641]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/ansible-kubernetes.d/nopull.yml owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1736612965.2957008-8350-171865569281770/.source.yml _original_basename=.uofn7p8f follow=False checksum=d5dc917e3cae36de03aa971a17ac473f86fdf934 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:29:26 managed-node2 python3.12[12772]: ansible-containers.podman.podman_play Invoked with state=created kube_file=/etc/containers/ansible-kubernetes.d/nopull.yml executable=podman annotation=None kube_file_content=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None debug=None quiet=None recreate=None userns=None log_level=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None Jan 11 11:29:26 managed-node2 systemd[1]: var-lib-containers-storage-overlay-compat2222882315-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay-compat2222882315-merged.mount has successfully entered the 'dead' state. Jan 11 11:29:26 managed-node2 rsyslogd[886]: imjournal: journal files changed, reloading... [v8.2412.0-1.el10 try https://www.rsyslog.com/e/0 ] Jan 11 11:29:26 managed-node2 kernel: evm: overlay not supported Jan 11 11:29:26 managed-node2 podman[12779]: 2025-01-11 11:29:26.527506046 -0500 EST m=+0.076751034 system refresh Jan 11 11:29:26 managed-node2 systemd[1]: var-lib-containers-storage-overlay-metacopy\x2dcheck928390053-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay-metacopy\x2dcheck928390053-merged.mount has successfully entered the 'dead' state. Jan 11 11:29:26 managed-node2 podman[12779]: 2025-01-11 11:29:26.802195054 -0500 EST m=+0.351440103 image build f41b26d639b72cab2f92dd19cf0a7b368c0addd834cde4dfbe6ba9cc0b5e9755 Jan 11 11:29:26 managed-node2 systemd[1]: Created slice machine.slice - Slice /machine. ░░ Subject: A start job for unit machine.slice has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit machine.slice has finished successfully. ░░ ░░ The job identifier is 1776. Jan 11 11:29:26 managed-node2 systemd[1]: Created slice machine-libpod_pod_8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844.slice - cgroup machine-libpod_pod_8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844.slice. ░░ Subject: A start job for unit machine-libpod_pod_8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844.slice has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit machine-libpod_pod_8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844.slice has finished successfully. ░░ ░░ The job identifier is 1775. Jan 11 11:29:26 managed-node2 podman[12779]: 2025-01-11 11:29:26.84857707 -0500 EST m=+0.397821985 container create 8b291853c44d64b81d1ceed7ffd8c7c0a03d32d84fe9fcd377032ab3ba163ed6 (image=localhost/podman-pause:5.3.1-1733097600, name=8f21b98d6977-infra, pod_id=8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844, io.buildah.version=1.38.0) Jan 11 11:29:26 managed-node2 podman[12779]: 2025-01-11 11:29:26.854692046 -0500 EST m=+0.403936935 pod create 8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844 (image=, name=nopull) Jan 11 11:29:28 managed-node2 podman[12779]: 2025-01-11 11:29:28.907614035 -0500 EST m=+2.456859238 image pull 9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f quay.io/libpod/testimage:20210610 Jan 11 11:29:28 managed-node2 podman[12779]: 2025-01-11 11:29:28.930557469 -0500 EST m=+2.479802384 container create 010c205b58090946e480f7d230e44e257dd9db66fbcfe0636fb2c68d69f1bba4 (image=quay.io/libpod/testimage:20210610, name=nopull-nopull, pod_id=8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844, created_by=test/system/build-testimage, io.buildah.version=1.21.0, app=test, io.containers.autoupdate=registry, created_at=2021-06-10T18:55:36Z) Jan 11 11:29:28 managed-node2 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jan 11 11:29:31 managed-node2 python3.12[13112]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jan 11 11:29:32 managed-node2 python3.12[13249]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:29:35 managed-node2 python3.12[13382]: ansible-ansible.legacy.dnf Invoked with name=['firewalld'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:29:36 managed-node2 python3.12[13514]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False scope=system no_block=False state=None enabled=None force=None Jan 11 11:29:36 managed-node2 python3.12[13647]: ansible-ansible.legacy.systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Jan 11 11:29:37 managed-node2 python3.12[13780]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['15001-15003/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Jan 11 11:29:39 managed-node2 python3.12[13911]: ansible-ansible.legacy.dnf Invoked with name=['python3-libselinux', 'python3-policycoreutils'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:29:40 managed-node2 python3.12[14043]: ansible-ansible.legacy.dnf Invoked with name=['policycoreutils-python-utils'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:29:41 managed-node2 python3.12[14175]: ansible-setup Invoked with filter=['ansible_selinux'] gather_subset=['all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Jan 11 11:29:42 managed-node2 python3.12[14335]: ansible-fedora.linux_system_roles.local_seport Invoked with ports=['15001-15003'] proto=tcp setype=http_port_t state=present local=False ignore_selinux_state=False reload=True Jan 11 11:29:43 managed-node2 python3.12[14466]: ansible-fedora.linux_system_roles.selinux_modules_facts Invoked Jan 11 11:29:47 managed-node2 python3.12[14597]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:29:49 managed-node2 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jan 11 11:29:49 managed-node2 podman[14738]: 2025-01-11 11:29:49.434170482 -0500 EST m=+0.252936003 image pull-error quay.io/linux-system-roles/this_is_a_bogus_image:latest initializing source docker://quay.io/linux-system-roles/this_is_a_bogus_image:latest: reading manifest latest in quay.io/linux-system-roles/this_is_a_bogus_image: unauthorized: access to the requested resource is not authorized Jan 11 11:29:49 managed-node2 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jan 11 11:29:49 managed-node2 python3.12[14876]: ansible-stat Invoked with path=/etc/containers/ansible-kubernetes.d/bogus.yml follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:29:50 managed-node2 python3.12[15007]: ansible-file Invoked with path=/etc/containers/ansible-kubernetes.d state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:29:50 managed-node2 python3.12[15138]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/ansible-kubernetes.d/bogus.yml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jan 11 11:29:50 managed-node2 python3.12[15243]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/ansible-kubernetes.d/bogus.yml owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1736612990.3622415-9267-128608785517959/.source.yml _original_basename=.lo9fbot2 follow=False checksum=f8266a972ed3be7e204d2a67883fe3a22b8dbf18 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:29:51 managed-node2 python3.12[15374]: ansible-containers.podman.podman_play Invoked with state=created kube_file=/etc/containers/ansible-kubernetes.d/bogus.yml executable=podman annotation=None kube_file_content=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None debug=None quiet=None recreate=None userns=None log_level=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None Jan 11 11:29:51 managed-node2 systemd[1]: Created slice machine-libpod_pod_6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7.slice - cgroup machine-libpod_pod_6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7.slice. ░░ Subject: A start job for unit machine-libpod_pod_6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7.slice has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit machine-libpod_pod_6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7.slice has finished successfully. ░░ ░░ The job identifier is 1781. Jan 11 11:29:51 managed-node2 podman[15381]: 2025-01-11 11:29:51.566032778 -0500 EST m=+0.059424724 container create 5a92effc85487f4e0d67a204339e8c66d39fd2acb0e397ba7a06494670bdc5fa (image=localhost/podman-pause:5.3.1-1733097600, name=6939fbfdbe78-infra, pod_id=6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7, io.buildah.version=1.38.0) Jan 11 11:29:51 managed-node2 podman[15381]: 2025-01-11 11:29:51.57210719 -0500 EST m=+0.065499244 pod create 6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7 (image=, name=bogus) Jan 11 11:29:52 managed-node2 podman[15381]: 2025-01-11 11:29:52.780406701 -0500 EST m=+1.273798703 image pull-error quay.io/linux-system-roles/this_is_a_bogus_image:latest initializing source docker://quay.io/linux-system-roles/this_is_a_bogus_image:latest: reading manifest latest in quay.io/linux-system-roles/this_is_a_bogus_image: unauthorized: access to the requested resource is not authorized Jan 11 11:29:52 managed-node2 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jan 11 11:29:55 managed-node2 python3.12[15650]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jan 11 11:29:56 managed-node2 python3.12[15787]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:29:58 managed-node2 python3.12[15920]: ansible-ansible.legacy.dnf Invoked with name=['firewalld'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:30:00 managed-node2 python3.12[16052]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False scope=system no_block=False state=None enabled=None force=None Jan 11 11:30:00 managed-node2 python3.12[16185]: ansible-ansible.legacy.systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Jan 11 11:30:01 managed-node2 python3.12[16318]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['15001-15003/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Jan 11 11:30:03 managed-node2 python3.12[16449]: ansible-ansible.legacy.dnf Invoked with name=['python3-libselinux', 'python3-policycoreutils'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:30:03 managed-node2 python3.12[16581]: ansible-ansible.legacy.dnf Invoked with name=['policycoreutils-python-utils'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:30:04 managed-node2 python3.12[16713]: ansible-setup Invoked with filter=['ansible_selinux'] gather_subset=['all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Jan 11 11:30:06 managed-node2 python3.12[16873]: ansible-fedora.linux_system_roles.local_seport Invoked with ports=['15001-15003'] proto=tcp setype=http_port_t state=present local=False ignore_selinux_state=False reload=True Jan 11 11:30:06 managed-node2 python3.12[17004]: ansible-fedora.linux_system_roles.selinux_modules_facts Invoked Jan 11 11:30:11 managed-node2 python3.12[17135]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:30:12 managed-node2 python3.12[17268]: ansible-ansible.legacy.command Invoked with _raw_params=systemd-escape --template podman-kube@.service /etc/containers/ansible-kubernetes.d/nopull.yml _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jan 11 11:30:13 managed-node2 python3.12[17400]: ansible-systemd Invoked with name=podman-kube@-etc-containers-ansible\x2dkubernetes.d-nopull.yml.service scope=system state=stopped enabled=False daemon_reload=False daemon_reexec=False no_block=False force=None masked=None Jan 11 11:30:13 managed-node2 python3.12[17533]: ansible-stat Invoked with path=/etc/containers/ansible-kubernetes.d/nopull.yml follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:30:14 managed-node2 python3.12[17666]: ansible-containers.podman.podman_play Invoked with state=absent kube_file=/etc/containers/ansible-kubernetes.d/nopull.yml executable=podman annotation=None kube_file_content=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None debug=None quiet=None recreate=None userns=None log_level=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None Jan 11 11:30:14 managed-node2 python3.12[17666]: ansible-containers.podman.podman_play version: 5.3.1, kube file /etc/containers/ansible-kubernetes.d/nopull.yml Jan 11 11:30:14 managed-node2 podman[17673]: 2025-01-11 11:30:14.120968075 -0500 EST m=+0.024014018 pod stop 8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844 (image=, name=nopull) Jan 11 11:30:14 managed-node2 systemd[1]: Removed slice machine-libpod_pod_8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844.slice - cgroup machine-libpod_pod_8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844.slice. ░░ Subject: A stop job for unit machine-libpod_pod_8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844.slice has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit machine-libpod_pod_8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844.slice has finished. ░░ ░░ The job identifier is 1787 and the job result is done. Jan 11 11:30:14 managed-node2 podman[17673]: 2025-01-11 11:30:14.1635176 -0500 EST m=+0.066563524 container remove 010c205b58090946e480f7d230e44e257dd9db66fbcfe0636fb2c68d69f1bba4 (image=quay.io/libpod/testimage:20210610, name=nopull-nopull, pod_id=8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, app=test) Jan 11 11:30:14 managed-node2 podman[17673]: 2025-01-11 11:30:14.18894132 -0500 EST m=+0.091987362 container remove 8b291853c44d64b81d1ceed7ffd8c7c0a03d32d84fe9fcd377032ab3ba163ed6 (image=localhost/podman-pause:5.3.1-1733097600, name=8f21b98d6977-infra, pod_id=8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844, io.buildah.version=1.38.0) Jan 11 11:30:14 managed-node2 podman[17673]: 2025-01-11 11:30:14.198654504 -0500 EST m=+0.101700390 pod remove 8f21b98d69777a1218bb5004f4328b2c021f4c67e942421bf447537e489b9844 (image=, name=nopull) Jan 11 11:30:14 managed-node2 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jan 11 11:30:14 managed-node2 python3.12[17813]: ansible-file Invoked with path=/etc/containers/ansible-kubernetes.d/nopull.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:30:15 managed-node2 python3.12[17944]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jan 11 11:30:15 managed-node2 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jan 11 11:30:18 managed-node2 python3.12[18214]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jan 11 11:30:19 managed-node2 python3.12[18351]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:30:22 managed-node2 python3.12[18484]: ansible-ansible.legacy.dnf Invoked with name=['firewalld'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:30:23 managed-node2 python3.12[18616]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False scope=system no_block=False state=None enabled=None force=None Jan 11 11:30:24 managed-node2 python3.12[18749]: ansible-ansible.legacy.systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Jan 11 11:30:24 managed-node2 python3.12[18882]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['15001-15003/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Jan 11 11:30:26 managed-node2 python3.12[19013]: ansible-ansible.legacy.dnf Invoked with name=['python3-libselinux', 'python3-policycoreutils'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:30:27 managed-node2 python3.12[19145]: ansible-ansible.legacy.dnf Invoked with name=['policycoreutils-python-utils'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:30:28 managed-node2 python3.12[19277]: ansible-setup Invoked with filter=['ansible_selinux'] gather_subset=['all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Jan 11 11:30:30 managed-node2 python3.12[19437]: ansible-fedora.linux_system_roles.local_seport Invoked with ports=['15001-15003'] proto=tcp setype=http_port_t state=present local=False ignore_selinux_state=False reload=True Jan 11 11:30:30 managed-node2 python3.12[19568]: ansible-fedora.linux_system_roles.selinux_modules_facts Invoked Jan 11 11:30:35 managed-node2 python3.12[19699]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:30:36 managed-node2 python3.12[19832]: ansible-ansible.legacy.command Invoked with _raw_params=systemd-escape --template podman-kube@.service /etc/containers/ansible-kubernetes.d/bogus.yml _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jan 11 11:30:37 managed-node2 python3.12[19964]: ansible-systemd Invoked with name=podman-kube@-etc-containers-ansible\x2dkubernetes.d-bogus.yml.service scope=system state=stopped enabled=False daemon_reload=False daemon_reexec=False no_block=False force=None masked=None Jan 11 11:30:38 managed-node2 python3.12[20097]: ansible-stat Invoked with path=/etc/containers/ansible-kubernetes.d/bogus.yml follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:30:38 managed-node2 python3.12[20230]: ansible-containers.podman.podman_play Invoked with state=absent kube_file=/etc/containers/ansible-kubernetes.d/bogus.yml executable=podman annotation=None kube_file_content=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None debug=None quiet=None recreate=None userns=None log_level=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None Jan 11 11:30:38 managed-node2 python3.12[20230]: ansible-containers.podman.podman_play version: 5.3.1, kube file /etc/containers/ansible-kubernetes.d/bogus.yml Jan 11 11:30:38 managed-node2 podman[20237]: 2025-01-11 11:30:38.640117526 -0500 EST m=+0.024351398 pod stop 6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7 (image=, name=bogus) Jan 11 11:30:38 managed-node2 systemd[1]: Removed slice machine-libpod_pod_6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7.slice - cgroup machine-libpod_pod_6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7.slice. ░░ Subject: A stop job for unit machine-libpod_pod_6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7.slice has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit machine-libpod_pod_6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7.slice has finished. ░░ ░░ The job identifier is 1789 and the job result is done. Jan 11 11:30:38 managed-node2 podman[20237]: 2025-01-11 11:30:38.680455077 -0500 EST m=+0.064688863 container remove 5a92effc85487f4e0d67a204339e8c66d39fd2acb0e397ba7a06494670bdc5fa (image=localhost/podman-pause:5.3.1-1733097600, name=6939fbfdbe78-infra, pod_id=6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7, io.buildah.version=1.38.0) Jan 11 11:30:38 managed-node2 podman[20237]: 2025-01-11 11:30:38.690033298 -0500 EST m=+0.074267051 pod remove 6939fbfdbe7806915cedaa414a8bc98c9b016a05e789d967c53a3b35b3cd92a7 (image=, name=bogus) Jan 11 11:30:38 managed-node2 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jan 11 11:30:39 managed-node2 python3.12[20377]: ansible-file Invoked with path=/etc/containers/ansible-kubernetes.d/bogus.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:30:39 managed-node2 python3.12[20508]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jan 11 11:30:39 managed-node2 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jan 11 11:30:42 managed-node2 python3.12[20778]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jan 11 11:30:43 managed-node2 python3.12[20916]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:30:47 managed-node2 python3.12[21049]: ansible-ansible.legacy.dnf Invoked with name=['firewalld'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:30:48 managed-node2 python3.12[21181]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False scope=system no_block=False state=None enabled=None force=None Jan 11 11:30:48 managed-node2 python3.12[21314]: ansible-ansible.legacy.systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Jan 11 11:30:49 managed-node2 python3.12[21447]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['15001-15003/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Jan 11 11:30:51 managed-node2 python3.12[21578]: ansible-ansible.legacy.dnf Invoked with name=['python3-libselinux', 'python3-policycoreutils'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:30:52 managed-node2 python3.12[21710]: ansible-ansible.legacy.dnf Invoked with name=['policycoreutils-python-utils'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Jan 11 11:30:53 managed-node2 python3.12[21842]: ansible-setup Invoked with filter=['ansible_selinux'] gather_subset=['all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Jan 11 11:30:54 managed-node2 python3.12[22002]: ansible-fedora.linux_system_roles.local_seport Invoked with ports=['15001-15003'] proto=tcp setype=http_port_t state=present local=False ignore_selinux_state=False reload=True Jan 11 11:30:55 managed-node2 python3.12[22133]: ansible-fedora.linux_system_roles.selinux_modules_facts Invoked Jan 11 11:30:59 managed-node2 python3.12[22264]: ansible-getent Invoked with database=passwd key=podman_basic_user fail_key=False service=None split=None Jan 11 11:31:00 managed-node2 python3.12[22396]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:31:00 managed-node2 python3.12[22529]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids podman_basic_user _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jan 11 11:31:01 managed-node2 python3.12[22661]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g podman_basic_user _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jan 11 11:31:02 managed-node2 python3.12[22793]: ansible-ansible.legacy.command Invoked with _raw_params=systemd-escape --template podman-kube@.service /home/podman_basic_user/.config/containers/ansible-kubernetes.d/httpd1.yml _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jan 11 11:31:03 managed-node2 python3.12[22925]: ansible-ansible.legacy.command Invoked with creates=/var/lib/systemd/linger/podman_basic_user _raw_params=loginctl enable-linger podman_basic_user _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None removes=None stdin=None Jan 11 11:31:03 managed-node2 systemd[1]: Created slice user-3001.slice - User Slice of UID 3001. ░░ Subject: A start job for unit user-3001.slice has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-3001.slice has finished successfully. ░░ ░░ The job identifier is 1869. Jan 11 11:31:03 managed-node2 systemd[1]: Starting user-runtime-dir@3001.service - User Runtime Directory /run/user/3001... ░░ Subject: A start job for unit user-runtime-dir@3001.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-runtime-dir@3001.service has begun execution. ░░ ░░ The job identifier is 1791. Jan 11 11:31:03 managed-node2 systemd[1]: Finished user-runtime-dir@3001.service - User Runtime Directory /run/user/3001. ░░ Subject: A start job for unit user-runtime-dir@3001.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-runtime-dir@3001.service has finished successfully. ░░ ░░ The job identifier is 1791. Jan 11 11:31:03 managed-node2 systemd[1]: Starting user@3001.service - User Manager for UID 3001... ░░ Subject: A start job for unit user@3001.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user@3001.service has begun execution. ░░ ░░ The job identifier is 1871. Jan 11 11:31:03 managed-node2 systemd-logind[659]: New session 6 of user podman_basic_user. ░░ Subject: A new session 6 has been created for user podman_basic_user ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 6 has been created for the user podman_basic_user. ░░ ░░ The leading process of the session is 22929. Jan 11 11:31:03 managed-node2 (systemd)[22929]: pam_unix(systemd-user:session): session opened for user podman_basic_user(uid=3001) by podman_basic_user(uid=0) Jan 11 11:31:03 managed-node2 systemd[22929]: Queued start job for default target default.target. Jan 11 11:31:03 managed-node2 systemd[22929]: Created slice app.slice - User Application Slice. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 5. Jan 11 11:31:03 managed-node2 systemd[22929]: Started grub-boot-success.timer - Mark boot as successful after the user session has run 2 minutes. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 11. Jan 11 11:31:03 managed-node2 systemd[22929]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of User's Temporary Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 12. Jan 11 11:31:03 managed-node2 systemd[22929]: Reached target paths.target - Paths. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 8. Jan 11 11:31:03 managed-node2 systemd[22929]: Reached target timers.target - Timers. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 10. Jan 11 11:31:03 managed-node2 systemd[22929]: Starting dbus.socket - D-Bus User Message Bus Socket... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 4. Jan 11 11:31:03 managed-node2 systemd[22929]: Starting systemd-tmpfiles-setup.service - Create User Files and Directories... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 9. Jan 11 11:31:03 managed-node2 systemd[22929]: Finished systemd-tmpfiles-setup.service - Create User Files and Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 9. Jan 11 11:31:03 managed-node2 systemd[22929]: Listening on dbus.socket - D-Bus User Message Bus Socket. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 4. Jan 11 11:31:03 managed-node2 systemd[22929]: Reached target sockets.target - Sockets. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 3. Jan 11 11:31:03 managed-node2 systemd[22929]: Reached target basic.target - Basic System. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 2. Jan 11 11:31:03 managed-node2 systemd[22929]: Reached target default.target - Main User Target. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 1. Jan 11 11:31:03 managed-node2 systemd[1]: Started user@3001.service - User Manager for UID 3001. ░░ Subject: A start job for unit user@3001.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user@3001.service has finished successfully. ░░ ░░ The job identifier is 1871. Jan 11 11:31:03 managed-node2 systemd[22929]: Startup finished in 69ms. ░░ Subject: User manager start-up is now complete ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The user manager instance for user 3001 has been started. All services queued ░░ for starting have been started. Note that other services might still be starting ░░ up or be started at any later time. ░░ ░░ Startup of the manager took 69160 microseconds. Jan 11 11:31:04 managed-node2 python3.12[23075]: ansible-file Invoked with path=/tmp/lsr_adhv0_3l_podman/httpd1 state=directory owner=podman_basic_user group=3001 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:31:04 managed-node2 python3.12[23206]: ansible-file Invoked with path=/tmp/lsr_adhv0_3l_podman/httpd1-create state=directory owner=podman_basic_user group=3001 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:31:05 managed-node2 sudo[23379]: root : TTY=pts/0 ; PWD=/root ; USER=podman_basic_user ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ebjivediwrynnjycvbbcvkpnrvrgvyty ; XDG_RUNTIME_DIR=/run/user/3001 /usr/bin/python3.12 /var/tmp/ansible-tmp-1736613064.7957578-12329-92755468246573/AnsiballZ_podman_image.py' Jan 11 11:31:05 managed-node2 sudo[23379]: pam_unix(sudo:session): session opened for user podman_basic_user(uid=3001) by root(uid=0) Jan 11 11:31:05 managed-node2 systemd[22929]: Created slice session.slice - User Core Session Slice. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 14. Jan 11 11:31:05 managed-node2 systemd[22929]: Starting dbus-broker.service - D-Bus User Message Bus... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 13. Jan 11 11:31:05 managed-node2 dbus-broker-launch[23403]: Policy to allow eavesdropping in /usr/share/dbus-1/session.conf +31: Eavesdropping is deprecated and ignored Jan 11 11:31:05 managed-node2 dbus-broker-launch[23403]: Policy to allow eavesdropping in /usr/share/dbus-1/session.conf +33: Eavesdropping is deprecated and ignored Jan 11 11:31:05 managed-node2 systemd[22929]: Started dbus-broker.service - D-Bus User Message Bus. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 13. Jan 11 11:31:05 managed-node2 dbus-broker-launch[23403]: Ready Jan 11 11:31:05 managed-node2 systemd[22929]: Created slice user.slice - Slice /user. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 20. Jan 11 11:31:05 managed-node2 systemd[22929]: Started podman-23389.scope. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 19. Jan 11 11:31:05 managed-node2 systemd[22929]: Started podman-pause-de278718.scope. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 23. Jan 11 11:31:05 managed-node2 systemd[22929]: Started podman-23405.scope. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 27. Jan 11 11:31:06 managed-node2 systemd[22929]: Started podman-23430.scope. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 31. Jan 11 11:31:06 managed-node2 sudo[23379]: pam_unix(sudo:session): session closed for user podman_basic_user Jan 11 11:31:07 managed-node2 python3.12[23567]: ansible-stat Invoked with path=/home/podman_basic_user/.config/containers/ansible-kubernetes.d/httpd1.yml follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jan 11 11:31:07 managed-node2 python3.12[23698]: ansible-file Invoked with path=/home/podman_basic_user/.config/containers/ansible-kubernetes.d state=directory owner=podman_basic_user group=3001 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:31:08 managed-node2 python3.12[23829]: ansible-ansible.legacy.stat Invoked with path=/home/podman_basic_user/.config/containers/ansible-kubernetes.d/httpd1.yml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jan 11 11:31:08 managed-node2 python3.12[23934]: ansible-ansible.legacy.copy Invoked with dest=/home/podman_basic_user/.config/containers/ansible-kubernetes.d/httpd1.yml owner=podman_basic_user group=3001 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1736613067.8332987-12432-130581053166951/.source.yml _original_basename=.vntwb0uw follow=False checksum=bd406dc7744755fdf41f83e27a5ef8497bec46ba backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Jan 11 11:31:08 managed-node2 sudo[24107]: root : TTY=pts/0 ; PWD=/root ; USER=podman_basic_user ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-hdtiduuiveohnpfnsmzuxmnranaqsupd ; XDG_RUNTIME_DIR=/run/user/3001 /usr/bin/python3.12 /var/tmp/ansible-tmp-1736613068.54518-12462-228328251565918/AnsiballZ_podman_play.py' Jan 11 11:31:08 managed-node2 sudo[24107]: pam_unix(sudo:session): session opened for user podman_basic_user(uid=3001) by root(uid=0) Jan 11 11:31:08 managed-node2 python3.12[24110]: ansible-containers.podman.podman_play Invoked with state=started debug=True log_level=debug kube_file=/home/podman_basic_user/.config/containers/ansible-kubernetes.d/httpd1.yml executable=podman annotation=None kube_file_content=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None quiet=None recreate=None userns=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None Jan 11 11:31:08 managed-node2 systemd[22929]: Started podman-24117.scope. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 35. Jan 11 11:31:09 managed-node2 systemd[22929]: Created slice user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice - cgroup user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 39. Jan 11 11:31:09 managed-node2 kernel: tun: Universal TUN/TAP device driver, 1.6 Jan 11 11:31:09 managed-node2 systemd[22929]: Started rootless-netns-832a999c.scope. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 43. Jan 11 11:31:09 managed-node2 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 11 11:31:09 managed-node2 kernel: podman1: port 1(veth0) entered blocking state Jan 11 11:31:09 managed-node2 kernel: podman1: port 1(veth0) entered disabled state Jan 11 11:31:09 managed-node2 kernel: veth0: entered allmulticast mode Jan 11 11:31:09 managed-node2 kernel: veth0: entered promiscuous mode Jan 11 11:31:09 managed-node2 kernel: podman1: port 1(veth0) entered blocking state Jan 11 11:31:09 managed-node2 kernel: podman1: port 1(veth0) entered forwarding state Jan 11 11:31:09 managed-node2 systemd[22929]: Started run-p24198-i24498.scope - [systemd-run] /usr/libexec/podman/aardvark-dns --config /run/user/3001/containers/networks/aardvark-dns -p 53 run. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 47. Jan 11 11:31:09 managed-node2 aardvark-dns[24198]: starting aardvark on a child with pid 24199 Jan 11 11:31:09 managed-node2 aardvark-dns[24199]: Successfully parsed config Jan 11 11:31:09 managed-node2 aardvark-dns[24199]: Listen v4 ip {"podman-default-kube-network": [10.89.0.1]} Jan 11 11:31:09 managed-node2 aardvark-dns[24199]: Listen v6 ip {} Jan 11 11:31:09 managed-node2 aardvark-dns[24199]: Using the following upstream servers: [169.254.1.1:53, 10.29.169.13:53, 10.29.170.12:53] Jan 11 11:31:09 managed-node2 conmon[24214]: conmon 6494e2ff6f9e6c9552ee : failed to write to /proc/self/oom_score_adj: Permission denied Jan 11 11:31:09 managed-node2 systemd[22929]: Started libpod-conmon-6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e.scope. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 51. Jan 11 11:31:09 managed-node2 conmon[24215]: conmon 6494e2ff6f9e6c9552ee : addr{sun_family=AF_UNIX, sun_path=/proc/self/fd/14/attach} Jan 11 11:31:09 managed-node2 conmon[24215]: conmon 6494e2ff6f9e6c9552ee : terminal_ctrl_fd: 14 Jan 11 11:31:09 managed-node2 conmon[24215]: conmon 6494e2ff6f9e6c9552ee : winsz read side: 17, winsz write side: 18 Jan 11 11:31:09 managed-node2 systemd[22929]: Started libpod-6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e.scope - libcrun container. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 56. Jan 11 11:31:09 managed-node2 conmon[24215]: conmon 6494e2ff6f9e6c9552ee : container PID: 24217 Jan 11 11:31:09 managed-node2 conmon[24215]: conmon 6494e2ff6f9e6c9552ee : container 24217 exited with status 127 Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Called cleanup.PersistentPreRunE(/usr/bin/podman --root /home/podman_basic_user/.local/share/containers/storage --runroot /run/user/3001/containers --log-level debug --cgroup-manager systemd --tmpdir /run/user/3001/libpod/tmp --network-config-dir --network-backend netavark --volumepath /home/podman_basic_user/.local/share/containers/storage/volumes --db-backend sqlite --transient-store=false --runtime crun --storage-driver overlay --events-backend file --syslog container cleanup --stopped-only 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e)" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Setting custom database backend: \"sqlite\"" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Using conmon: \"/usr/bin/conmon\"" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=info msg="Using sqlite as database backend" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="systemd-logind: Unknown object '/'." Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Using graph driver overlay" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Using graph root /home/podman_basic_user/.local/share/containers/storage" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Using run root /run/user/3001/containers" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Using static dir /home/podman_basic_user/.local/share/containers/storage/libpod" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Using tmp dir /run/user/3001/libpod/tmp" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Using volume path /home/podman_basic_user/.local/share/containers/storage/volumes" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Using transient store: false" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="[graphdriver] trying provided driver \"overlay\"" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Cached value indicated that overlay is supported" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Cached value indicated that overlay is supported" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Cached value indicated that metacopy is not being used" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Cached value indicated that native-diff is usable" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="backingFs=xfs, projectQuotaSupported=false, useNativeDiff=true, usingMetacopy=false" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Initializing event backend file" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Configured OCI runtime runj initialization failed: no valid executable found for OCI runtime runj: invalid argument" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Configured OCI runtime youki initialization failed: no valid executable found for OCI runtime youki: invalid argument" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Configured OCI runtime crun-vm initialization failed: no valid executable found for OCI runtime crun-vm: invalid argument" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Configured OCI runtime runc initialization failed: no valid executable found for OCI runtime runc: invalid argument" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Configured OCI runtime kata initialization failed: no valid executable found for OCI runtime kata: invalid argument" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Configured OCI runtime runsc initialization failed: no valid executable found for OCI runtime runsc: invalid argument" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Configured OCI runtime krun initialization failed: no valid executable found for OCI runtime krun: invalid argument" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Configured OCI runtime ocijail initialization failed: no valid executable found for OCI runtime ocijail: invalid argument" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Configured OCI runtime crun-wasm initialization failed: no valid executable found for OCI runtime crun-wasm: invalid argument" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Using OCI runtime \"/usr/bin/crun\"" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=info msg="Setting parallel job count to 7" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Cleaning up container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Tearing down network namespace at /run/user/3001/netns/netns-1ba1aabb-9ce7-ed09-ced0-c13ed359f037 for container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Successfully loaded network podman-default-kube-network: &{podman-default-kube-network fb42127ceae097e95afb4c22c535aaa88e07cb9147b8dd0d6c01dc2c233a4302 bridge podman1 2025-01-11 11:31:08.992158363 -0500 EST [{{{10.89.0.0 ffffff00}} 10.89.0.1 }] [] false false true [] map[] map[] map[driver:host-local]}" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Successfully loaded 2 networks" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="The path of /etc/resolv.conf in the mount ns is \"/etc/resolv.conf\"" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=info msg="netavark: [DEBUG netavark::commands::teardown] Tearing down..\n" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=info msg="netavark: [INFO netavark::firewall] Using nftables firewall driver\n" Jan 11 11:31:09 managed-node2 aardvark-dns[24199]: Received SIGHUP Jan 11 11:31:09 managed-node2 aardvark-dns[24199]: Successfully parsed config Jan 11 11:31:09 managed-node2 aardvark-dns[24199]: Listen v4 ip {} Jan 11 11:31:09 managed-node2 aardvark-dns[24199]: Listen v6 ip {} Jan 11 11:31:09 managed-node2 aardvark-dns[24199]: No configuration found stopping the sever Jan 11 11:31:09 managed-node2 kernel: podman1: port 1(veth0) entered disabled state Jan 11 11:31:09 managed-node2 kernel: veth0 (unregistering): left allmulticast mode Jan 11 11:31:09 managed-node2 kernel: veth0 (unregistering): left promiscuous mode Jan 11 11:31:09 managed-node2 kernel: podman1: port 1(veth0) entered disabled state Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=info msg="netavark: [INFO netavark::network::bridge] removing bridge podman1\n" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=info msg="netavark: [DEBUG netavark::firewall::nft] Matched Rule { family: INet, table: \"netavark\", chain: \"INPUT\", expr: [Match(Match { left: Named(Payload(PayloadField(PayloadField { protocol: \"ip\", field: \"saddr\" }))), right: Named(Prefix(Prefix { addr: String(\"10.89.0.0\"), len: 24 })), op: EQ }), Match(Match { left: Named(Meta(Meta { key: L4proto })), right: Named(Set([Element(String(\"tcp\")), Element(String(\"udp\"))])), op: EQ }), Match(Match { left: Named(Payload(PayloadField(PayloadField { protocol: \"th\", field: \"dport\" }))), right: Number(53), op: EQ }), Accept(None)], handle: Some(23), index: None, comment: None }\n[DEBUG netavark::firewall::nft] Matched Rule { family: INet, table: \"netavark\", chain: \"FORWARD\", expr: [Match(Match { left: Named(Payload(PayloadField(PayloadField { protocol: \"ip\", field: \"daddr\" }))), right: Named(Prefix(Prefix { addr: String(\"10.89.0.0\"), len: 24 })), op: EQ }), Match(Match { left: Named(CT(CT { key: \"state\", family: None, dir: None })), right: List([String(\"established\"), String(\"related\")]), op: IN }), Accept(None)], handle: Some(24), index: None, comment: None }\n" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=info msg="netavark: [DEBUG netavark::firewall::nft] Matched Rule { family: INet, table: \"netavark\", chain: \"FORWARD\", expr: [Match(Match { left: Named(Payload(PayloadField(PayloadField { protocol: \"ip\", field: \"saddr\" }))), right: Named(Prefix(Prefix { addr: String(\"10.89.0.0\"), len: 24 })), op: EQ }), Accept(None)], handle: Some(25), index: None, comment: None }\n[DEBUG netavark::firewall::nft] Matched Rule { family: INet, table: \"netavark\", chain: \"POSTROUTING\", expr: [Match(Match { left: Named(Payload(PayloadField(PayloadField { protocol: \"ip\", field: \"saddr\" }))), right: Named(Prefix(Prefix { addr: String(\"10.89.0.0\"), len: 24 })), op: EQ }), Jump(JumpTarget { target: \"nv_fb42127c_10_89_0_0_nm24\" })], handle: Some(26), index: None, comment: None }\n[DEBUG netavark::firewall::nft] Removing 4 rules\n[DEBUG netavark::firewall::nft] Found chain nv_fb42127c_10_89_0_0_nm24\n" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=info msg="netavark: [DEBUG netavark::firewall::nft] Matched Rule { family: INet, table: \"netavark\", chain: \"NETAVARK-ISOLATION-3\", expr: [Match(Match { left: Named(Meta(Meta { key: Oifname })), right: String(\"podman1\"), op: EQ }), Drop(None)], handle: Some(17), index: None, comment: None }\n[DEBUG netavark::firewall::nft] Removing 1 isolation rules for network\n" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=info msg="netavark: [DEBUG netavark::firewall::nft] Found chain nv_fb42127c_10_89_0_0_nm24_dnat\n[DEBUG netavark::firewall::nft] Found chain nv_fb42127c_10_89_0_0_nm24_dnat\n" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=info msg="netavark: [DEBUG netavark::commands::teardown] Teardown complete\n" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Cleaning up rootless network namespace" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Successfully cleaned up container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Unmounted container \"6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e\"" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Called cleanup.PersistentPostRunE(/usr/bin/podman --root /home/podman_basic_user/.local/share/containers/storage --runroot /run/user/3001/containers --log-level debug --cgroup-manager systemd --tmpdir /run/user/3001/libpod/tmp --network-config-dir --network-backend netavark --volumepath /home/podman_basic_user/.local/share/containers/storage/volumes --db-backend sqlite --transient-store=false --runtime crun --storage-driver overlay --events-backend file --syslog container cleanup --stopped-only 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e)" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=debug msg="Shutting down engines" Jan 11 11:31:09 managed-node2 /usr/bin/podman[24219]: time="2025-01-11T11:31:09-05:00" level=info msg="Received shutdown.Stop(), terminating!" PID=24219 Jan 11 11:31:09 managed-node2 systemd[22929]: Stopping libpod-conmon-6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e.scope... ░░ Subject: A stop job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit UNIT has begun execution. ░░ ░░ The job identifier is 62. Jan 11 11:31:09 managed-node2 systemd[22929]: Stopped libpod-conmon-6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e.scope. ░░ Subject: A stop job for unit UNIT has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit UNIT has finished. ░░ ░░ The job identifier is 62 and the job result is done. Jan 11 11:31:09 managed-node2 systemd[22929]: Removed slice user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice - cgroup user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice. ░░ Subject: A stop job for unit UNIT has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit UNIT has finished. ░░ ░░ The job identifier is 61 and the job result is done. Jan 11 11:31:09 managed-node2 python3.12[24110]: ansible-containers.podman.podman_play PODMAN-PLAY-KUBE command: /bin/podman play kube --start=true --log-level=debug /home/podman_basic_user/.config/containers/ansible-kubernetes.d/httpd1.yml Jan 11 11:31:09 managed-node2 python3.12[24110]: ansible-containers.podman.podman_play PODMAN-PLAY-KUBE stdout: [starting container 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de: cannot get namespace path unless container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e is running: container is stopped] Pod: cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da Container: 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de Jan 11 11:31:09 managed-node2 python3.12[24110]: ansible-containers.podman.podman_play PODMAN-PLAY-KUBE stderr: time="2025-01-11T11:31:08-05:00" level=info msg="/bin/podman filtering at log level debug" time="2025-01-11T11:31:08-05:00" level=debug msg="Called kube.PersistentPreRunE(/bin/podman play kube --start=true --log-level=debug /home/podman_basic_user/.config/containers/ansible-kubernetes.d/httpd1.yml)" time="2025-01-11T11:31:08-05:00" level=debug msg="Using conmon: \"/usr/bin/conmon\"" time="2025-01-11T11:31:08-05:00" level=info msg="Using sqlite as database backend" time="2025-01-11T11:31:08-05:00" level=debug msg="systemd-logind: Unknown object '/'." time="2025-01-11T11:31:08-05:00" level=debug msg="Using graph driver overlay" time="2025-01-11T11:31:08-05:00" level=debug msg="Using graph root /home/podman_basic_user/.local/share/containers/storage" time="2025-01-11T11:31:08-05:00" level=debug msg="Using run root /run/user/3001/containers" time="2025-01-11T11:31:08-05:00" level=debug msg="Using static dir /home/podman_basic_user/.local/share/containers/storage/libpod" time="2025-01-11T11:31:08-05:00" level=debug msg="Using tmp dir /run/user/3001/libpod/tmp" time="2025-01-11T11:31:08-05:00" level=debug msg="Using volume path /home/podman_basic_user/.local/share/containers/storage/volumes" time="2025-01-11T11:31:08-05:00" level=debug msg="Using transient store: false" time="2025-01-11T11:31:08-05:00" level=debug msg="[graphdriver] trying provided driver \"overlay\"" time="2025-01-11T11:31:08-05:00" level=debug msg="Cached value indicated that overlay is supported" time="2025-01-11T11:31:08-05:00" level=debug msg="Cached value indicated that overlay is supported" time="2025-01-11T11:31:08-05:00" level=debug msg="Cached value indicated that metacopy is not being used" time="2025-01-11T11:31:08-05:00" level=debug msg="Cached value indicated that native-diff is usable" time="2025-01-11T11:31:08-05:00" level=debug msg="backingFs=xfs, projectQuotaSupported=false, useNativeDiff=true, usingMetacopy=false" time="2025-01-11T11:31:08-05:00" level=debug msg="Initializing event backend file" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime ocijail initialization failed: no valid executable found for OCI runtime ocijail: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime crun-vm initialization failed: no valid executable found for OCI runtime crun-vm: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime youki initialization failed: no valid executable found for OCI runtime youki: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime kata initialization failed: no valid executable found for OCI runtime kata: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime runsc initialization failed: no valid executable found for OCI runtime runsc: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime krun initialization failed: no valid executable found for OCI runtime krun: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime crun-wasm initialization failed: no valid executable found for OCI runtime crun-wasm: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime runc initialization failed: no valid executable found for OCI runtime runc: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Configured OCI runtime runj initialization failed: no valid executable found for OCI runtime runj: invalid argument" time="2025-01-11T11:31:08-05:00" level=debug msg="Using OCI runtime \"/usr/bin/crun\"" time="2025-01-11T11:31:08-05:00" level=info msg="Setting parallel job count to 7" time="2025-01-11T11:31:08-05:00" level=debug msg="Successfully loaded 1 networks" time="2025-01-11T11:31:08-05:00" level=debug msg="found free device name podman1" time="2025-01-11T11:31:08-05:00" level=debug msg="found free ipv4 network subnet 10.89.0.0/24" time="2025-01-11T11:31:08-05:00" level=debug msg="Looking up image \"localhost/podman-pause:5.3.1-1733097600\" in local containers storage" time="2025-01-11T11:31:08-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:08-05:00" level=debug msg="Trying \"localhost/podman-pause:5.3.1-1733097600\" ..." time="2025-01-11T11:31:08-05:00" level=debug msg="reference \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]localhost/podman-pause:5.3.1-1733097600\" does not resolve to an image ID" time="2025-01-11T11:31:08-05:00" level=debug msg="Trying \"localhost/podman-pause:5.3.1-1733097600\" ..." time="2025-01-11T11:31:08-05:00" level=debug msg="reference \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]localhost/podman-pause:5.3.1-1733097600\" does not resolve to an image ID" time="2025-01-11T11:31:08-05:00" level=debug msg="Trying \"localhost/podman-pause:5.3.1-1733097600\" ..." time="2025-01-11T11:31:08-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:08-05:00" level=debug msg="FROM \"scratch\"" time="2025-01-11T11:31:08-05:00" level=debug msg="Cached value indicated that idmapped mounts for overlay are not supported" time="2025-01-11T11:31:08-05:00" level=debug msg="Check for idmapped mounts support " time="2025-01-11T11:31:08-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:08-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:09-05:00" level=debug msg="overlay: test mount indicated that volatile is being used" time="2025-01-11T11:31:09-05:00" level=debug msg="overlay: mount_data=lowerdir=/home/podman_basic_user/.local/share/containers/storage/overlay/ce5d425d1a334b7a751d21348811c909114800550f7ebdee36d06daf3bf588d0/empty,upperdir=/home/podman_basic_user/.local/share/containers/storage/overlay/ce5d425d1a334b7a751d21348811c909114800550f7ebdee36d06daf3bf588d0/diff,workdir=/home/podman_basic_user/.local/share/containers/storage/overlay/ce5d425d1a334b7a751d21348811c909114800550f7ebdee36d06daf3bf588d0/work,userxattr,volatile,context=\"system_u:object_r:container_file_t:s0:c681,c972\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Container ID: 108a4b73ea294c92e350e6914ba70ae43bc510d0757a34527239170a2fe2fa22" time="2025-01-11T11:31:09-05:00" level=debug msg="Parsed Step: {Env:[PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin] Command:copy Args:[/usr/libexec/podman/catatonit /catatonit] Flags:[] Attrs:map[] Message:COPY /usr/libexec/podman/catatonit /catatonit Heredocs:[] Original:COPY /usr/libexec/podman/catatonit /catatonit}" time="2025-01-11T11:31:09-05:00" level=debug msg="COPY []string(nil), imagebuilder.Copy{FromFS:false, From:\"\", Src:[]string{\"/usr/libexec/podman/catatonit\"}, Dest:\"/catatonit\", Download:false, Chown:\"\", Chmod:\"\", Checksum:\"\", Files:[]imagebuilder.File(nil), KeepGitDir:false, Link:false, Parents:false, Excludes:[]string(nil)}" time="2025-01-11T11:31:09-05:00" level=debug msg="EnsureContainerPath \"/\" (owner \"\", mode 0) in \"108a4b73ea294c92e350e6914ba70ae43bc510d0757a34527239170a2fe2fa22\"" time="2025-01-11T11:31:09-05:00" level=debug msg="added content file:872045bc026632654869cebc14f91fe531969760bf26925e889d9b70527c8b67" time="2025-01-11T11:31:09-05:00" level=debug msg="Parsed Step: {Env:[PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin] Command:entrypoint Args:[/catatonit -P] Flags:[] Attrs:map[json:true] Message:ENTRYPOINT /catatonit -P Heredocs:[] Original:ENTRYPOINT [\"/catatonit\", \"-P\"]}" time="2025-01-11T11:31:09-05:00" level=debug msg="EnsureContainerPath \"/\" (owner \"\", mode 0) in \"108a4b73ea294c92e350e6914ba70ae43bc510d0757a34527239170a2fe2fa22\"" time="2025-01-11T11:31:09-05:00" level=debug msg="COMMIT localhost/podman-pause:5.3.1-1733097600" time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]localhost/podman-pause:5.3.1-1733097600\"" time="2025-01-11T11:31:09-05:00" level=debug msg="COMMIT \"containers-storage:[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]localhost/podman-pause:5.3.1-1733097600\"" time="2025-01-11T11:31:09-05:00" level=debug msg="committing image with reference \"containers-storage:[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]localhost/podman-pause:5.3.1-1733097600\" is allowed by policy" time="2025-01-11T11:31:09-05:00" level=debug msg="layer list: [\"ce5d425d1a334b7a751d21348811c909114800550f7ebdee36d06daf3bf588d0\"]" time="2025-01-11T11:31:09-05:00" level=debug msg="using \"/var/tmp/buildah4034309036\" to hold temporary data" time="2025-01-11T11:31:09-05:00" level=debug msg="Tar with options on /home/podman_basic_user/.local/share/containers/storage/overlay/ce5d425d1a334b7a751d21348811c909114800550f7ebdee36d06daf3bf588d0/diff" time="2025-01-11T11:31:09-05:00" level=debug msg="layer \"ce5d425d1a334b7a751d21348811c909114800550f7ebdee36d06daf3bf588d0\" size is 699392 bytes, uncompressed digest sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6, possibly-compressed digest sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6" time="2025-01-11T11:31:09-05:00" level=debug msg="OCIv1 config = {\"created\":\"2025-01-11T16:31:09.16093935Z\",\"architecture\":\"amd64\",\"os\":\"linux\",\"config\":{\"Env\":[\"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin\"],\"Entrypoint\":[\"/catatonit\",\"-P\"],\"WorkingDir\":\"/\",\"Labels\":{\"io.buildah.version\":\"1.38.0\"}},\"rootfs\":{\"type\":\"layers\",\"diff_ids\":[\"sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6\"]},\"history\":[{\"created\":\"2025-01-11T16:31:09.133319185Z\",\"created_by\":\"/bin/sh -c #(nop) COPY file:872045bc026632654869cebc14f91fe531969760bf26925e889d9b70527c8b67 in /catatonit \",\"empty_layer\":true},{\"created\":\"2025-01-11T16:31:09.164063409Z\",\"created_by\":\"/bin/sh -c #(nop) ENTRYPOINT [\\\"/catatonit\\\", \\\"-P\\\"]\"}]}" time="2025-01-11T11:31:09-05:00" level=debug msg="OCIv1 manifest = {\"schemaVersion\":2,\"mediaType\":\"application/vnd.oci.image.manifest.v1+json\",\"config\":{\"mediaType\":\"application/vnd.oci.image.config.v1+json\",\"digest\":\"sha256:5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\",\"size\":684},\"layers\":[{\"mediaType\":\"application/vnd.oci.image.layer.v1.tar\",\"digest\":\"sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6\",\"size\":699392}],\"annotations\":{\"org.opencontainers.image.base.digest\":\"\",\"org.opencontainers.image.base.name\":\"\"}}" time="2025-01-11T11:31:09-05:00" level=debug msg="Docker v2s2 config = {\"created\":\"2025-01-11T16:31:09.16093935Z\",\"container\":\"108a4b73ea294c92e350e6914ba70ae43bc510d0757a34527239170a2fe2fa22\",\"container_config\":{\"Hostname\":\"\",\"Domainname\":\"\",\"User\":\"\",\"AttachStdin\":false,\"AttachStdout\":false,\"AttachStderr\":false,\"Tty\":false,\"OpenStdin\":false,\"StdinOnce\":false,\"Env\":[\"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin\"],\"Cmd\":null,\"Image\":\"\",\"Volumes\":{},\"WorkingDir\":\"/\",\"Entrypoint\":[\"/catatonit\",\"-P\"],\"OnBuild\":[],\"Labels\":{\"io.buildah.version\":\"1.38.0\"}},\"config\":{\"Hostname\":\"\",\"Domainname\":\"\",\"User\":\"\",\"AttachStdin\":false,\"AttachStdout\":false,\"AttachStderr\":false,\"Tty\":false,\"OpenStdin\":false,\"StdinOnce\":false,\"Env\":[\"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin\"],\"Cmd\":null,\"Image\":\"\",\"Volumes\":{},\"WorkingDir\":\"/\",\"Entrypoint\":[\"/catatonit\",\"-P\"],\"OnBuild\":[],\"Labels\":{\"io.buildah.version\":\"1.38.0\"}},\"architecture\":\"amd64\",\"os\":\"linux\",\"rootfs\":{\"type\":\"layers\",\"diff_ids\":[\"sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6\"]},\"history\":[{\"created\":\"2025-01-11T16:31:09.133319185Z\",\"created_by\":\"/bin/sh -c #(nop) COPY file:872045bc026632654869cebc14f91fe531969760bf26925e889d9b70527c8b67 in /catatonit \",\"empty_layer\":true},{\"created\":\"2025-01-11T16:31:09.164063409Z\",\"created_by\":\"/bin/sh -c #(nop) ENTRYPOINT [\\\"/catatonit\\\", \\\"-P\\\"]\"}]}" time="2025-01-11T11:31:09-05:00" level=debug msg="Docker v2s2 manifest = {\"schemaVersion\":2,\"mediaType\":\"application/vnd.docker.distribution.manifest.v2+json\",\"config\":{\"mediaType\":\"application/vnd.docker.container.image.v1+json\",\"size\":1347,\"digest\":\"sha256:3d5363d5b8025414d8363b8279ab1974a53f63e86ef685536caca8cc6f1f8c3a\"},\"layers\":[{\"mediaType\":\"application/vnd.docker.image.rootfs.diff.tar\",\"size\":699392,\"digest\":\"sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6\"}]}" time="2025-01-11T11:31:09-05:00" level=debug msg="Using SQLite blob info cache at /home/podman_basic_user/.local/share/containers/cache/blob-info-cache-v1.sqlite" time="2025-01-11T11:31:09-05:00" level=debug msg="IsRunningImageAllowed for image containers-storage:" time="2025-01-11T11:31:09-05:00" level=debug msg=" Using transport \"containers-storage\" policy section \"\"" time="2025-01-11T11:31:09-05:00" level=debug msg=" Requirement 0: allowed" time="2025-01-11T11:31:09-05:00" level=debug msg="Overall: allowed" time="2025-01-11T11:31:09-05:00" level=debug msg="start reading config" time="2025-01-11T11:31:09-05:00" level=debug msg="finished reading config" time="2025-01-11T11:31:09-05:00" level=debug msg="Manifest has MIME type application/vnd.oci.image.manifest.v1+json, ordered candidate list [application/vnd.oci.image.manifest.v1+json, application/vnd.docker.distribution.manifest.v2+json, application/vnd.docker.distribution.manifest.v1+prettyjws, application/vnd.docker.distribution.manifest.v1+json]" time="2025-01-11T11:31:09-05:00" level=debug msg="... will first try using the original manifest unmodified" time="2025-01-11T11:31:09-05:00" level=debug msg="Checking if we can reuse blob sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6: general substitution = true, compression for MIME type \"application/vnd.oci.image.layer.v1.tar\" = true" time="2025-01-11T11:31:09-05:00" level=debug msg="reading layer \"sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6\"" time="2025-01-11T11:31:09-05:00" level=debug msg="No compression detected" time="2025-01-11T11:31:09-05:00" level=debug msg="Using original blob without modification" time="2025-01-11T11:31:09-05:00" level=debug msg="Applying tar in /home/podman_basic_user/.local/share/containers/storage/overlay/83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6/diff" time="2025-01-11T11:31:09-05:00" level=debug msg="finished reading layer \"sha256:83362ec78cab9ea69b1fd7cce8c7ba3164dab292c6df41c7dbcb6971d43d89e6\"" time="2025-01-11T11:31:09-05:00" level=debug msg="No compression detected" time="2025-01-11T11:31:09-05:00" level=debug msg="Compression change for blob sha256:5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49 (\"application/vnd.oci.image.config.v1+json\") not supported" time="2025-01-11T11:31:09-05:00" level=debug msg="Using original blob without modification" time="2025-01-11T11:31:09-05:00" level=debug msg="setting image creation date to 2025-01-11 16:31:09.16093935 +0000 UTC" time="2025-01-11T11:31:09-05:00" level=debug msg="created new image ID \"5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\" with metadata \"{}\"" time="2025-01-11T11:31:09-05:00" level=debug msg="added name \"localhost/podman-pause:5.3.1-1733097600\" to image \"5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\"" time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]localhost/podman-pause:5.3.1-1733097600\"" time="2025-01-11T11:31:09-05:00" level=debug msg="printing final image id \"5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Pod using bridge network mode" time="2025-01-11T11:31:09-05:00" level=debug msg="Created cgroup path user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice for parent user.slice and name libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da" time="2025-01-11T11:31:09-05:00" level=debug msg="Created cgroup user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice" time="2025-01-11T11:31:09-05:00" level=debug msg="Got pod cgroup as user.slice/user-3001.slice/user@3001.service/user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice" time="2025-01-11T11:31:09-05:00" level=debug msg="Looking up image \"localhost/podman-pause:5.3.1-1733097600\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:09-05:00" level=debug msg="Trying \"localhost/podman-pause:5.3.1-1733097600\" ..." time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"localhost/podman-pause:5.3.1-1733097600\" as \"localhost/podman-pause:5.3.1-1733097600\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"localhost/podman-pause:5.3.1-1733097600\" as \"localhost/podman-pause:5.3.1-1733097600\" in local containers storage ([overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49)" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49" time="2025-01-11T11:31:09-05:00" level=debug msg="using systemd mode: false" time="2025-01-11T11:31:09-05:00" level=debug msg="setting container name cec703044d41-infra" time="2025-01-11T11:31:09-05:00" level=debug msg="Loading seccomp profile from \"/usr/share/containers/seccomp.json\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Successfully loaded network podman-default-kube-network: &{podman-default-kube-network fb42127ceae097e95afb4c22c535aaa88e07cb9147b8dd0d6c01dc2c233a4302 bridge podman1 2025-01-11 11:31:08.992158363 -0500 EST [{{{10.89.0.0 ffffff00}} 10.89.0.1 }] [] false false true [] map[] map[] map[driver:host-local]}" time="2025-01-11T11:31:09-05:00" level=debug msg="Successfully loaded 2 networks" time="2025-01-11T11:31:09-05:00" level=debug msg="Allocated lock 1 for container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:5e29eea82466556190590f3bfa54c36cf146f7444d2d419150ffcc40f2526d49\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Created container \"6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Container \"6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e\" has work directory \"/home/podman_basic_user/.local/share/containers/storage/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Container \"6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e\" has run directory \"/run/user/3001/containers/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Looking up image \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:09-05:00" level=debug msg="Trying \"quay.io/libpod/testimage:20210610\" ..." time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage ([overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f)" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Pulling image quay.io/libpod/testimage:20210610 (policy: missing)" time="2025-01-11T11:31:09-05:00" level=debug msg="Looking up image \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:09-05:00" level=debug msg="Trying \"quay.io/libpod/testimage:20210610\" ..." time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage ([overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f)" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Looking up image \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:09-05:00" level=debug msg="Trying \"quay.io/libpod/testimage:20210610\" ..." time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage ([overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f)" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Looking up image \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Normalized platform linux/amd64 to {amd64 linux [] }" time="2025-01-11T11:31:09-05:00" level=debug msg="Trying \"quay.io/libpod/testimage:20210610\" ..." time="2025-01-11T11:31:09-05:00" level=debug msg="parsed reference into \"[overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage" time="2025-01-11T11:31:09-05:00" level=debug msg="Found image \"quay.io/libpod/testimage:20210610\" as \"quay.io/libpod/testimage:20210610\" in local containers storage ([overlay@/home/podman_basic_user/.local/share/containers/storage+/run/user/3001/containers]@9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f)" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f" time="2025-01-11T11:31:09-05:00" level=debug msg="Inspecting image 9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f" time="2025-01-11T11:31:09-05:00" level=debug msg="using systemd mode: false" time="2025-01-11T11:31:09-05:00" level=debug msg="adding container to pod httpd1" time="2025-01-11T11:31:09-05:00" level=debug msg="setting container name httpd1-httpd1" time="2025-01-11T11:31:09-05:00" level=debug msg="Loading seccomp profile from \"/usr/share/containers/seccomp.json\"" time="2025-01-11T11:31:09-05:00" level=info msg="Sysctl net.ipv4.ping_group_range=0 0 ignored in containers.conf, since Network Namespace set to host" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding mount /proc" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding mount /dev" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding mount /dev/pts" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding mount /dev/mqueue" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding mount /sys" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding mount /sys/fs/cgroup" time="2025-01-11T11:31:09-05:00" level=debug msg="Allocated lock 2 for container 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de" time="2025-01-11T11:31:09-05:00" level=debug msg="exporting opaque data as blob \"sha256:9f9ec7f2fdef9168f74e9d057f307955db14d782cff22ded51d277d74798cb2f\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Created container \"69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Container \"69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de\" has work directory \"/home/podman_basic_user/.local/share/containers/storage/overlay-containers/69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de/userdata\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Container \"69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de\" has run directory \"/run/user/3001/containers/overlay-containers/69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de/userdata\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Strongconnecting node 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" time="2025-01-11T11:31:09-05:00" level=debug msg="Pushed 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e onto stack" time="2025-01-11T11:31:09-05:00" level=debug msg="Finishing node 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e. Popped 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e off stack" time="2025-01-11T11:31:09-05:00" level=debug msg="Strongconnecting node 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de" time="2025-01-11T11:31:09-05:00" level=debug msg="Pushed 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de onto stack" time="2025-01-11T11:31:09-05:00" level=debug msg="Finishing node 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de. Popped 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de off stack" time="2025-01-11T11:31:09-05:00" level=debug msg="overlay: mount_data=lowerdir=/home/podman_basic_user/.local/share/containers/storage/overlay/l/VY45JJLRBH5N22CXKVWPXHFS6U,upperdir=/home/podman_basic_user/.local/share/containers/storage/overlay/438cf6e5d85031b39f6e2286fb90558bcb520d90dfcaf76539b1082e932b7376/diff,workdir=/home/podman_basic_user/.local/share/containers/storage/overlay/438cf6e5d85031b39f6e2286fb90558bcb520d90dfcaf76539b1082e932b7376/work,userxattr,context=\"system_u:object_r:container_file_t:s0:c471,c842\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Made network namespace at /run/user/3001/netns/netns-1ba1aabb-9ce7-ed09-ced0-c13ed359f037 for container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" time="2025-01-11T11:31:09-05:00" level=debug msg="Mounted container \"6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e\" at \"/home/podman_basic_user/.local/share/containers/storage/overlay/438cf6e5d85031b39f6e2286fb90558bcb520d90dfcaf76539b1082e932b7376/merged\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Created root filesystem for container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e at /home/podman_basic_user/.local/share/containers/storage/overlay/438cf6e5d85031b39f6e2286fb90558bcb520d90dfcaf76539b1082e932b7376/merged" time="2025-01-11T11:31:09-05:00" level=debug msg="Creating rootless network namespace at \"/run/user/3001/containers/networks/rootless-netns/rootless-netns\"" time="2025-01-11T11:31:09-05:00" level=debug msg="pasta arguments: --config-net --pid /run/user/3001/containers/networks/rootless-netns/rootless-netns-conn.pid --dns-forward 169.254.1.1 -t none -u none -T none -U none --no-map-gw --quiet --netns /run/user/3001/containers/networks/rootless-netns/rootless-netns --map-guest-addr 169.254.1.2" time="2025-01-11T11:31:09-05:00" level=debug msg="The path of /etc/resolv.conf in the mount ns is \"/etc/resolv.conf\"" [DEBUG netavark::network::validation] Validating network namespace... [DEBUG netavark::commands::setup] Setting up... [INFO netavark::firewall] Using nftables firewall driver [DEBUG netavark::network::bridge] Setup network podman-default-kube-network [DEBUG netavark::network::bridge] Container interface name: eth0 with IP addresses [10.89.0.2/24] [DEBUG netavark::network::bridge] Bridge name: podman1 with IP addresses [10.89.0.1/24] [DEBUG netavark::network::core_utils] Setting sysctl value for net.ipv4.ip_forward to 1 [DEBUG netavark::network::core_utils] Setting sysctl value for /proc/sys/net/ipv4/conf/podman1/rp_filter to 2 [DEBUG netavark::network::core_utils] Setting sysctl value for /proc/sys/net/ipv6/conf/eth0/autoconf to 0 [DEBUG netavark::network::core_utils] Setting sysctl value for /proc/sys/net/ipv4/conf/eth0/arp_notify to 1 [DEBUG netavark::network::core_utils] Setting sysctl value for /proc/sys/net/ipv4/conf/eth0/rp_filter to 2 [INFO netavark::network::netlink] Adding route (dest: 0.0.0.0/0 ,gw: 10.89.0.1, metric 100) [INFO netavark::firewall::nft] Creating container chain nv_fb42127c_10_89_0_0_nm24 [DEBUG netavark::network::core_utils] Setting sysctl value for net.ipv4.conf.podman1.route_localnet to 1 [DEBUG netavark::dns::aardvark] Spawning aardvark server [DEBUG netavark::dns::aardvark] start aardvark-dns: ["systemd-run", "-q", "--scope", "--user", "/usr/libexec/podman/aardvark-dns", "--config", "/run/user/3001/containers/networks/aardvark-dns", "-p", "53", "run"] [DEBUG netavark::commands::setup] { "podman-default-kube-network": StatusBlock { dns_search_domains: Some( [ "dns.podman", ], ), dns_server_ips: Some( [ 10.89.0.1, ], ), interfaces: Some( { "eth0": NetInterface { mac_address: "ba:4e:01:fd:9a:cc", subnets: Some( [ NetAddress { gateway: Some( 10.89.0.1, ), ipnet: 10.89.0.2/24, }, ], ), }, }, ), }, } [DEBUG netavark::commands::setup] Setup complete time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport: time=\"2025-01-11T11:31:09-05:00\" level=info msg=\"Starting parent driver\"\n" time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport: time=\"2025-01-11T11:31:09-05:00\" level=info msg=\"opaque=map[builtin.readypipepath:/run/user/3001/libpod/tmp/rootlessport1778471301/.bp-ready.pipe builtin.socketpath:/run/user/3001/libpod/tmp/rootlessport1778471301/.bp.sock]\"\n" time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport: time=\"2025-01-11T11:31:09-05:00\" level=info msg=\"Starting child driver in child netns (\\\"/proc/self/exe\\\" [rootlessport-child])\"\n" time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport: time=\"2025-01-11T11:31:09-05:00\" level=info msg=\"Waiting for initComplete\"\n" time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport: time=\"2025-01-11T11:31:09-05:00\" level=info msg=\"initComplete is closed; parent and child established the communication channel\"\ntime=\"2025-01-11T11:31:09-05:00\" level=info msg=\"Exposing ports [{ 80 15001 1 tcp}]\"\n" time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport: time=\"2025-01-11T11:31:09-05:00\" level=info msg=Ready\n" time="2025-01-11T11:31:09-05:00" level=debug msg="rootlessport is ready" time="2025-01-11T11:31:09-05:00" level=debug msg="/proc/sys/crypto/fips_enabled does not contain '1', not adding FIPS mode bind mounts" time="2025-01-11T11:31:09-05:00" level=debug msg="Setting Cgroups for container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e to user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice:libpod:6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" time="2025-01-11T11:31:09-05:00" level=debug msg="reading hooks from /usr/share/containers/oci/hooks.d" time="2025-01-11T11:31:09-05:00" level=debug msg="Workdir \"/\" resolved to host path \"/home/podman_basic_user/.local/share/containers/storage/overlay/438cf6e5d85031b39f6e2286fb90558bcb520d90dfcaf76539b1082e932b7376/merged\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Created OCI spec for container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e at /home/podman_basic_user/.local/share/containers/storage/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata/config.json" time="2025-01-11T11:31:09-05:00" level=debug msg="Created cgroup path user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice for parent user.slice and name libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da" time="2025-01-11T11:31:09-05:00" level=debug msg="Created cgroup user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice" time="2025-01-11T11:31:09-05:00" level=debug msg="Got pod cgroup as user.slice/user-3001.slice/user@3001.service/user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice" time="2025-01-11T11:31:09-05:00" level=debug msg="/usr/bin/conmon messages will be logged to syslog" time="2025-01-11T11:31:09-05:00" level=debug msg="running conmon: /usr/bin/conmon" args="[--api-version 1 -c 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e -u 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e -r /usr/bin/crun -b /home/podman_basic_user/.local/share/containers/storage/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata -p /run/user/3001/containers/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata/pidfile -n cec703044d41-infra --exit-dir /run/user/3001/libpod/tmp/exits --persist-dir /run/user/3001/libpod/tmp/persist/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e --full-attach -s -l k8s-file:/home/podman_basic_user/.local/share/containers/storage/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata/ctr.log --log-level debug --syslog --conmon-pidfile /run/user/3001/containers/overlay-containers/6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e/userdata/conmon.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /home/podman_basic_user/.local/share/containers/storage --exit-command-arg --runroot --exit-command-arg /run/user/3001/containers --exit-command-arg --log-level --exit-command-arg debug --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/user/3001/libpod/tmp --exit-command-arg --network-config-dir --exit-command-arg --exit-command-arg --network-backend --exit-command-arg netavark --exit-command-arg --volumepath --exit-command-arg /home/podman_basic_user/.local/share/containers/storage/volumes --exit-command-arg --db-backend --exit-command-arg sqlite --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg crun --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --events-backend --exit-command-arg file --exit-command-arg --syslog --exit-command-arg container --exit-command-arg cleanup --exit-command-arg --stopped-only --exit-command-arg 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e]" time="2025-01-11T11:31:09-05:00" level=info msg="Running conmon under slice user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice and unitName libpod-conmon-6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e.scope" [conmon:d]: failed to write to /proc/self/oom_score_adj: Permission denied time="2025-01-11T11:31:09-05:00" level=debug msg="Received: 24217" time="2025-01-11T11:31:09-05:00" level=info msg="Got Conmon PID as 24215" time="2025-01-11T11:31:09-05:00" level=debug msg="Created container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e in OCI runtime" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding nameserver(s) from network status of '[\"10.89.0.1\"]'" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding search domain(s) from network status of '[\"dns.podman\"]'" time="2025-01-11T11:31:09-05:00" level=debug msg="Starting container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e with command [/catatonit -P]" time="2025-01-11T11:31:09-05:00" level=debug msg="Started container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" time="2025-01-11T11:31:09-05:00" level=debug msg="overlay: mount_data=lowerdir=/home/podman_basic_user/.local/share/containers/storage/overlay/l/H6MLQJZURTWK5UKPYM3NBCTJTC,upperdir=/home/podman_basic_user/.local/share/containers/storage/overlay/2fa690399eece911fb5509ec500fd7d2ca654be99b65c5d16219b201e4a935b1/diff,workdir=/home/podman_basic_user/.local/share/containers/storage/overlay/2fa690399eece911fb5509ec500fd7d2ca654be99b65c5d16219b201e4a935b1/work,userxattr,context=\"system_u:object_r:container_file_t:s0:c471,c842\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Mounted container \"69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de\" at \"/home/podman_basic_user/.local/share/containers/storage/overlay/2fa690399eece911fb5509ec500fd7d2ca654be99b65c5d16219b201e4a935b1/merged\"" time="2025-01-11T11:31:09-05:00" level=debug msg="Created root filesystem for container 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de at /home/podman_basic_user/.local/share/containers/storage/overlay/2fa690399eece911fb5509ec500fd7d2ca654be99b65c5d16219b201e4a935b1/merged" time="2025-01-11T11:31:09-05:00" level=debug msg="/proc/sys/crypto/fips_enabled does not contain '1', not adding FIPS mode bind mounts" time="2025-01-11T11:31:09-05:00" level=debug msg="Cleaning up container 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de" time="2025-01-11T11:31:09-05:00" level=debug msg="Unmounted container \"69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de\"" starting container 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de: cannot get namespace path unless container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e is running: container is stopped Error: failed to start 1 containers time="2025-01-11T11:31:09-05:00" level=debug msg="Shutting down engines" time="2025-01-11T11:31:09-05:00" level=info msg="Received shutdown.Stop(), terminating!" PID=24117 time="2025-01-11T11:31:09-05:00" level=debug msg="Adding parallel job to stop container 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e" time="2025-01-11T11:31:09-05:00" level=debug msg="Adding parallel job to stop container 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de" time="2025-01-11T11:31:09-05:00" level=debug msg="Stopping ctr 69655705528cd205d56c08a3a238862e03c6ed0ca137f32e8a037ee66e10b7de (timeout 10)" time="2025-01-11T11:31:09-05:00" level=debug msg="Stopping ctr 6494e2ff6f9e6c9552ee567ccbabb9c0a860226f6795e3fc26107d320571412e (timeout 10)" time="2025-01-11T11:31:09-05:00" level=debug msg="Removing pod cgroup user.slice/user-3001.slice/user@3001.service/user.slice/user-libpod_pod_cec703044d41b698edb70f2702fb0cd75b3c99d6fb620e924512450b4a7be9da.slice" Jan 11 11:31:09 managed-node2 python3.12[24110]: ansible-containers.podman.podman_play PODMAN-PLAY-KUBE rc: 125 Jan 11 11:31:09 managed-node2 sudo[24107]: pam_unix(sudo:session): session closed for user podman_basic_user Jan 11 11:31:10 managed-node2 python3.12[24369]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None TASK [Clean up storage.conf] *************************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:456 Saturday 11 January 2025 11:31:10 -0500 (0:00:00.528) 0:02:35.383 ****** ok: [managed-node2] => { "changed": false, "path": "/etc/containers/storage.conf", "state": "absent" } TASK [Clean up host directories] *********************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:463 Saturday 11 January 2025 11:31:10 -0500 (0:00:00.432) 0:02:35.815 ****** changed: [managed-node2] => { "changed": true, "path": "/tmp/lsr_adhv0_3l_podman", "state": "absent" } TASK [Remove kube file src] **************************************************** task path: /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:470 Saturday 11 January 2025 11:31:11 -0500 (0:00:00.391) 0:02:36.207 ****** changed: [managed-node2 -> localhost] => { "changed": true, "path": "/tmp/lsr_podman_yvb6_g1x.yml", "state": "absent" } PLAY RECAP ********************************************************************* managed-node2 : ok=283 changed=29 unreachable=0 failed=1 skipped=483 rescued=2 ignored=0 TASKS RECAP ******************************************************************** Saturday 11 January 2025 11:31:11 -0500 (0:00:00.283) 0:02:36.490 ****** =============================================================================== fedora.linux_system_roles.podman : Ensure required packages are installed -- 18.69s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 fedora.linux_system_roles.selinux : Get SELinux modules facts ----------- 3.32s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:112 fedora.linux_system_roles.selinux : Get SELinux modules facts ----------- 3.23s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:112 fedora.linux_system_roles.selinux : Get SELinux modules facts ----------- 3.07s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:112 fedora.linux_system_roles.podman : Update containers/pods --------------- 3.07s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:80 fedora.linux_system_roles.selinux : Get SELinux modules facts ----------- 3.06s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:112 fedora.linux_system_roles.selinux : Get SELinux modules facts ----------- 3.01s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:112 Gathering Facts --------------------------------------------------------- 2.45s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:9 fedora.linux_system_roles.podman : Ensure container images are present --- 2.25s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:29 Create data files ------------------------------------------------------- 2.00s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:204 fedora.linux_system_roles.podman : Update containers/pods --------------- 1.82s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:80 fedora.linux_system_roles.selinux : Set an SELinux label on a port ------ 1.41s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/main.yml:87 fedora.linux_system_roles.podman : Update containers/pods --------------- 1.37s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_kube_spec.yml:80 fedora.linux_system_roles.selinux : Install SELinux tool semanage ------- 1.19s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:58 Create host directories for data ---------------------------------------- 1.08s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/tests/podman/tests_basic.yml:196 fedora.linux_system_roles.firewall : Enable and start firewalld service --- 1.04s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 fedora.linux_system_roles.firewall : Install firewalld ------------------ 1.03s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 fedora.linux_system_roles.podman : Gather the package facts ------------- 1.01s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.podman : Gather the package facts ------------- 0.99s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.selinux : Refresh facts ----------------------- 0.95s /tmp/collections-BPh/ansible_collections/fedora/linux_system_roles/roles/selinux/tasks/ensure_selinux_packages.yml:89