ansible-playbook [core 2.17.7] config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.12/site-packages/ansible ansible collection location = /tmp/collections-9ny executable location = /usr/local/bin/ansible-playbook python version = 3.12.6 (main, Oct 29 2024, 00:00:00) [GCC 14.2.1 20240801 (Red Hat 14.2.1-2)] (/usr/bin/python3.12) jinja version = 3.1.4 libyaml = True No config file found; using defaults running playbook inside collection fedora.linux_system_roles redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug redirecting (type: callback) ansible.builtin.profile_tasks to ansible.posix.profile_tasks Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_quadlet_demo.yml *********************************************** 2 plays in /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml PLAY [all] ********************************************************************* TASK [Include vault variables] ************************************************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:5 Saturday 14 December 2024 11:32:50 -0500 (0:00:00.011) 0:00:00.011 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_test_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n35383939616163653333633431363463313831383037386236646138333162396161356130303461\n3932623930643263313563336163316337643562333936360a363538636631313039343233383732\n38666530383538656639363465313230343533386130303833336434303438333161656262346562\n3362626538613031640a663330613638366132356534363534353239616666653466353961323533\n6565\n" }, "mysql_container_root_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n61333932373230333539663035366431326163363166363036323963623131363530326231303634\n6635326161643165363366323062333334363730376631660a393566366139353861656364656661\n38653463363837336639363032646433666361646535366137303464623261313663643336306465\n6264663730656337310a343962353137386238383064646533366433333437303566656433386233\n34343235326665646661623131643335313236313131353661386338343366316261643634653633\n3832313034366536616531323963333234326461353130303532\n" } }, "ansible_included_var_files": [ "/tmp/podman-olA/tests/vars/vault-variables.yml" ], "changed": false } PLAY [Deploy the quadlet demo app] ********************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:9 Saturday 14 December 2024 11:32:51 -0500 (0:00:00.030) 0:00:00.041 ***** [WARNING]: Platform linux on host managed-node3 is using the discovered Python interpreter at /usr/bin/python3.12, but future installation of another Python interpreter could change the meaning of that path. See https://docs.ansible.com/ansible- core/2.17/reference_appendices/interpreter_discovery.html for more information. ok: [managed-node3] TASK [Test is only supported on x86_64] **************************************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:38 Saturday 14 December 2024 11:32:52 -0500 (0:00:01.482) 0:00:01.524 ***** skipping: [managed-node3] => { "false_condition": "ansible_facts[\"architecture\"] != \"x86_64\"" } TASK [End test] **************************************************************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:45 Saturday 14 December 2024 11:32:52 -0500 (0:00:00.034) 0:00:01.559 ***** META: end_play conditional evaluated to False, continuing play skipping: [managed-node3] => { "skip_reason": "end_play conditional evaluated to False, continuing play" } MSG: end_play TASK [Generate certificates] *************************************************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:51 Saturday 14 December 2024 11:32:52 -0500 (0:00:00.013) 0:00:01.572 ***** included: fedora.linux_system_roles.certificate for managed-node3 TASK [fedora.linux_system_roles.certificate : Set version specific variables] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:2 Saturday 14 December 2024 11:32:52 -0500 (0:00:00.111) 0:00:01.683 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.certificate : Ensure ansible_facts used by role] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:2 Saturday 14 December 2024 11:32:52 -0500 (0:00:00.049) 0:00:01.733 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__certificate_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.certificate : Check if system is ostree] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:10 Saturday 14 December 2024 11:32:52 -0500 (0:00:00.065) 0:00:01.799 ***** ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.certificate : Set flag to indicate system is ostree] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:15 Saturday 14 December 2024 11:32:53 -0500 (0:00:00.518) 0:00:02.317 ***** ok: [managed-node3] => { "ansible_facts": { "__certificate_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.certificate : Set platform/version specific variables] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:19 Saturday 14 December 2024 11:32:53 -0500 (0:00:00.028) 0:00:02.346 ***** skipping: [managed-node3] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node3] => (item=CentOS_10.yml) => { "ansible_facts": { "__certificate_certmonger_packages": [ "certmonger", "python3-packaging" ] }, "ansible_included_var_files": [ "/tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } ok: [managed-node3] => (item=CentOS_10.yml) => { "ansible_facts": { "__certificate_certmonger_packages": [ "certmonger", "python3-packaging" ] }, "ansible_included_var_files": [ "/tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } TASK [fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 Saturday 14 December 2024 11:32:53 -0500 (0:00:00.074) 0:00:02.420 ***** changed: [managed-node3] => { "changed": true, "rc": 0, "results": [ "Installed: python3-cffi-1.16.0-7.el10.x86_64", "Installed: python3-cryptography-43.0.0-4.el10.x86_64", "Installed: python3-ply-3.11-25.el10.noarch", "Installed: python3-pycparser-2.20-16.el10.noarch", "Installed: python3-pyasn1-0.6.1-1.el10.noarch" ] } lsrpackages: python3-cryptography python3-dbus python3-pyasn1 TASK [fedora.linux_system_roles.certificate : Ensure provider packages are installed] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 Saturday 14 December 2024 11:32:56 -0500 (0:00:02.625) 0:00:05.046 ***** changed: [managed-node3] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "rc": 0, "results": [ "Installed: certmonger-0.79.20-3.el10.x86_64", "Installed: python3-packaging-23.2-6.el10.noarch", "Installed: nspr-4.35.0-34.el10.x86_64", "Installed: nss-3.101.0-13.el10.x86_64", "Installed: dbus-tools-1:1.14.10-5.el10.x86_64", "Installed: nss-softokn-3.101.0-13.el10.x86_64", "Installed: nss-softokn-freebl-3.101.0-13.el10.x86_64", "Installed: nss-sysinit-3.101.0-13.el10.x86_64", "Installed: nss-util-3.101.0-13.el10.x86_64" ] } lsrpackages: certmonger python3-packaging TASK [fedora.linux_system_roles.certificate : Ensure pre-scripts hooks directory exists] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:35 Saturday 14 December 2024 11:32:59 -0500 (0:00:03.056) 0:00:08.102 ***** changed: [managed-node3] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//pre-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure post-scripts hooks directory exists] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:61 Saturday 14 December 2024 11:32:59 -0500 (0:00:00.569) 0:00:08.672 ***** changed: [managed-node3] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//post-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure provider service is running] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 Saturday 14 December 2024 11:33:00 -0500 (0:00:00.446) 0:00:09.118 ***** changed: [managed-node3] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": true, "enabled": true, "name": "certmonger", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:certmonger_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "syslog.target dbus-broker.service sysinit.target basic.target system.slice network.target systemd-journald.socket dbus.socket", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedorahosted.certmonger", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "Certificate monitoring and PKI enrollment", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveMemoryHigh": "3698237440", "EffectiveMemoryMax": "3698237440", "EffectiveTasksMax": "22361", "EnvironmentFiles": "/etc/sysconfig/certmonger (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/certmonger.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "certmonger.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3236470784", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "[not set]", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "[not set]", "MemorySwapMax": "infinity", "MemorySwapPeak": "[not set]", "MemoryZSwapCurrent": "[not set]", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "certmonger.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "PIDFile": "/run/certmonger.pid", "PartOf": "dbus-broker.service", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target dbus.socket system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22361", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "disabled", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.certificate : Ensure certificate requests] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101 Saturday 14 December 2024 11:33:01 -0500 (0:00:01.274) 0:00:10.392 ***** changed: [managed-node3] => (item={'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}) => { "ansible_loop_var": "item", "changed": true, "item": { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } } MSG: Certificate requested (new). TASK [fedora.linux_system_roles.certificate : Slurp the contents of the files] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152 Saturday 14 December 2024 11:33:02 -0500 (0:00:00.911) 0:00:11.303 ***** ok: [managed-node3] => (item=['cert', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnakNDQW1xZ0F3SUJBZ0lRVGZZOTdpbzJSUTJQQ2lKVXE4L3NKakFOQmdrcWhraUc5dzBCQVFzRkFEQlEKTVNBd0hnWURWUVFEREJkTWIyTmhiQ0JUYVdkdWFXNW5JRUYxZEdodmNtbDBlVEVzTUNvR0ExVUVBd3dqTkdSbQpOak5rWldVdE1tRXpOalExTUdRdE9HWXdZVEl5TlRRdFlXSmpabVZqTWpVd0hoY05NalF4TWpFME1UWXpNekF5CldoY05NalV4TWpFME1UWXpNekF4V2pBVU1SSXdFQVlEVlFRREV3bHNiMk5oYkdodmMzUXdnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRREtKWldVYkZFdHRHcjFMR2ZIUHVQcHJhM0d4UTN1MEtmRgpPRnZFWExwUkRLd1RwL21hNXFjeVRVTW03RHlad2MvcTk2bHl2ZmpuOEphVVYvU0JydkZ0STl0S003UmN1YlFJCktMWVI4TEpyNDE0clpXdHFqQnF6ZDVjUjFKQmkwK0lkMDl6eDh4NkZVbHN3K2V5WENaUExYZ1QxQXA4Sy9zVUsKeE5BTWwwS2hGcVhCMm1KelI4bmZnTTVOaHZtRCtySmNGZ2ViWTBlOHV6SUZBZlRvbjg1MkRrQzNtekdFZWF0dApjQ0ZWWnBDdlBrZk5CTXZJZjZ2dEprRmVoTWVKNnZPeWcxRUhZV1VWV1J3TmNGSHRmNURzY2NtTGdXWlJlREloCjArS0ZYdVJMYjYrSWl3dzUwOS9ldWVPMytnMjhJREpzdlZTcTR6emtjd2VvQnNWV1lqV3RBZ01CQUFHamdaTXcKZ1pBd0N3WURWUjBQQkFRREFnV2dNQlFHQTFVZEVRUU5NQXVDQ1d4dlkyRnNhRzl6ZERBZEJnTlZIU1VFRmpBVQpCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3REFZRFZSMFRBUUgvQkFJd0FEQWRCZ05WSFE0RUZnUVUzaXJRCjRPOUppY2RmbnNERWxndmdPS0ZxTkJVd0h3WURWUjBqQkJnd0ZvQVVhQWxaYlNxUy9ZUUtUdVBBSWRtb1daUTkKMmN3d0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFENEZXeGo5STBJMWNRN3VhRnptTmFvMVhKczlyd3c3ZEhHZApkSGpNcE1LTkphcFkzUTVyMUV3L2xSbzFNQkErdEhWR0tMWE1LL05qRTlNdU84dUdibzR6dVZxL2Z1RDZtVG1oCmJMTEZ0bCthcFJtM2xwcXdxUVR1WHZUdDlFbW9XRVhoTG01ME0rT0lNeUpuclJnbTlaWTVwbnhDZDJ6TDJycHoKUFdudUkyNXlDRnN5cUVrSWlKTUhnOHJJbWhROHdpZ2dOVDNjSlp0K1hNNis4bURHWDRCK3YrR3JsOEJEUVhwaQpMVVRrcFViWTJETTB6dlhOOWk2em9WQ2JoZUlpMnMzUzdZVmNsV21VY1BoMFpqRFJOVlpQbnhyRzdUeDBvMXlZCnpkdU5vdHhpSHBMNUFOSWRqTXhzUzlSYzdIKys1ekU0eG1GQzJlemlKWCtsWHVlZEVXVT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "cert", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } ok: [managed-node3] => (item=['key', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBQUklWQVRFIEtFWS0tLS0tCk1JSUV2Z0lCQURBTkJna3Foa2lHOXcwQkFRRUZBQVNDQktnd2dnU2tBZ0VBQW9JQkFRREtKWldVYkZFdHRHcjEKTEdmSFB1UHByYTNHeFEzdTBLZkZPRnZFWExwUkRLd1RwL21hNXFjeVRVTW03RHlad2MvcTk2bHl2ZmpuOEphVQpWL1NCcnZGdEk5dEtNN1JjdWJRSUtMWVI4TEpyNDE0clpXdHFqQnF6ZDVjUjFKQmkwK0lkMDl6eDh4NkZVbHN3CitleVhDWlBMWGdUMUFwOEsvc1VLeE5BTWwwS2hGcVhCMm1KelI4bmZnTTVOaHZtRCtySmNGZ2ViWTBlOHV6SUYKQWZUb244NTJEa0MzbXpHRWVhdHRjQ0ZWWnBDdlBrZk5CTXZJZjZ2dEprRmVoTWVKNnZPeWcxRUhZV1VWV1J3TgpjRkh0ZjVEc2NjbUxnV1pSZURJaDArS0ZYdVJMYjYrSWl3dzUwOS9ldWVPMytnMjhJREpzdlZTcTR6emtjd2VvCkJzVldZald0QWdNQkFBRUNnZ0VBRFpPSG43L0VsZmE3aGpyVXYzZDBWQ0hqZldtSG5ZWHE5cDVwMzE3QTdxZVUKdVA4N2p0RTE0blY0OEpqR0d5ZCtvTS9BVkNINUlRNng5a1dQK05PNGlEVnp5V0IrbVFtcDB6QjB4ZkN0OHZ3Qwp1NVJrcWZPOTZWYTdVU20vQnJ3blFBeVdGclBXMVdwcUVIaUFrTVpHWkU2c3M0M1pBNktIdGtqd2xoL2tqQy84Ckk0QVJzQ2JweGRQR2FocEM0N2lydXRVeHRhcElENHdrRzZTbWZ6bTg3M2FvZXd4WkRUVHVoMFF6cGcrdDRrci8KeFVtRi9RSWc1bkxsYU9YQ0hCOEl3VUxzd0tWVnZsbGRBMktVUnA0OE9NMUF1VDFPdFZKTGFVTGJvYnBKNSttTApiLys1M1ZBOUg1U2V0V0dXQitqc2d1Zm5RenlZSEVrMXdnMVk5cnh0d1FLQmdRRG5GYWJJQXZ1ZE94dk5HWGxCCmMzd21Ra1BucVphVG53UlFwRTFtWlR3UkkzMXVodndCSzI5L0w5NnVaYy95V05ycmRvMDRPc3dIcXlNMTFnNlQKRzFxZHhiSGR4SHB2YkgrNWlpbXk0UW1EWVBabHJvbmsvYlVvQWNFVGNzQlo2YWhsZUNRZWpndmsyTTA1VStVYwpKcEY3aHltMU1zN1hubk85REdHTVdHUTlUUUtCZ1FEZjhUTGRoQUl5VzIxMEJiQXg2akR1MUtRN2hZQ2t4WEViCkpTbFdlenUyK0p6ZEpEYlhKem1YR2dTT2RkWVRrSWNnYWd6ekRiMFJ0dGVUd3F3UCtLazhka1lJL3NvV3IxUDQKSHJ6TURma0RrV01NUGhoMC9VWmJGV0tWa3hxUUw3dWk5azZSQkVURlVqQ1UvTDFxUGRLZmhvdC80UU1oWUZOegpBK3g0V3RNcDRRS0JnUURjd050TjhVT0xKWjllWkhwb296a3psMVFsWitYWDZpNEdGNTkxTlRxWHFNNjNtcTM5Ck5CQlVjVmVLUk9yTk9rTVFjM1ZmQlJSS014djAzeU9YTnFkdlZ4K2Vtb2JZRU05SFlReUt2TjRJcXNWWHg5QWwKZUpUcHZodERSL2tuSzhiYUFGYUk5ZXU0MTcvZGh4NjBzTWVFT1ZBSHNKeldjS0ZpTlFTdlBQV3RkUUtCZ1FDSwpZaEFlbHA2TTQ3YUZSYWFWa2JSZlFZT2pUR0lzYjkzdStrM0xKU2ZIYjJYdnVac3g1UWxQZUVKc2ppQ1Z3bS9kCjllcStuelRBcDQ0YnZ1WjRrQjUwUUZvUzRPK2NpUWVJZ2VZRTZuV0FZTzBsK1pFL2FldDZDWUw4RHdVUktJMzQKcTkvcytRaE9MdlhhZXp2SlZSNjhlWFhtdGZNclFzME9lOVE2bzFWdHdRS0JnRlpPTDhaS0dGdVBwdDhlMGJlSwphdUVUVzRmSzN3eTBjWHlRaHJPZWN6ZGZkZEsrSzh4aTNWclZjUVhPRFQyb2Q0ajIzTnJKMEZiVEtVWFE1aEc2ClkvZGFwKytvbVJXa2xPWitXUE1kaXVlaEVCZDBLeXlMYWs4Z0FIVU5jZXIyQnV1SENobUw2TUl1L2o3N0tQTHIKVWJMQW53dm9EVlpMNTJKT3VUWk1FUWU3Ci0tLS0tRU5EIFBSSVZBVEUgS0VZLS0tLS0K", "encoding": "base64", "item": [ "key", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/private/quadlet_demo.key" } ok: [managed-node3] => (item=['ca', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnakNDQW1xZ0F3SUJBZ0lRVGZZOTdpbzJSUTJQQ2lKVXE4L3NKakFOQmdrcWhraUc5dzBCQVFzRkFEQlEKTVNBd0hnWURWUVFEREJkTWIyTmhiQ0JUYVdkdWFXNW5JRUYxZEdodmNtbDBlVEVzTUNvR0ExVUVBd3dqTkdSbQpOak5rWldVdE1tRXpOalExTUdRdE9HWXdZVEl5TlRRdFlXSmpabVZqTWpVd0hoY05NalF4TWpFME1UWXpNekF5CldoY05NalV4TWpFME1UWXpNekF4V2pBVU1SSXdFQVlEVlFRREV3bHNiMk5oYkdodmMzUXdnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRREtKWldVYkZFdHRHcjFMR2ZIUHVQcHJhM0d4UTN1MEtmRgpPRnZFWExwUkRLd1RwL21hNXFjeVRVTW03RHlad2MvcTk2bHl2ZmpuOEphVVYvU0JydkZ0STl0S003UmN1YlFJCktMWVI4TEpyNDE0clpXdHFqQnF6ZDVjUjFKQmkwK0lkMDl6eDh4NkZVbHN3K2V5WENaUExYZ1QxQXA4Sy9zVUsKeE5BTWwwS2hGcVhCMm1KelI4bmZnTTVOaHZtRCtySmNGZ2ViWTBlOHV6SUZBZlRvbjg1MkRrQzNtekdFZWF0dApjQ0ZWWnBDdlBrZk5CTXZJZjZ2dEprRmVoTWVKNnZPeWcxRUhZV1VWV1J3TmNGSHRmNURzY2NtTGdXWlJlREloCjArS0ZYdVJMYjYrSWl3dzUwOS9ldWVPMytnMjhJREpzdlZTcTR6emtjd2VvQnNWV1lqV3RBZ01CQUFHamdaTXcKZ1pBd0N3WURWUjBQQkFRREFnV2dNQlFHQTFVZEVRUU5NQXVDQ1d4dlkyRnNhRzl6ZERBZEJnTlZIU1VFRmpBVQpCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3REFZRFZSMFRBUUgvQkFJd0FEQWRCZ05WSFE0RUZnUVUzaXJRCjRPOUppY2RmbnNERWxndmdPS0ZxTkJVd0h3WURWUjBqQkJnd0ZvQVVhQWxaYlNxUy9ZUUtUdVBBSWRtb1daUTkKMmN3d0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFENEZXeGo5STBJMWNRN3VhRnptTmFvMVhKczlyd3c3ZEhHZApkSGpNcE1LTkphcFkzUTVyMUV3L2xSbzFNQkErdEhWR0tMWE1LL05qRTlNdU84dUdibzR6dVZxL2Z1RDZtVG1oCmJMTEZ0bCthcFJtM2xwcXdxUVR1WHZUdDlFbW9XRVhoTG01ME0rT0lNeUpuclJnbTlaWTVwbnhDZDJ6TDJycHoKUFdudUkyNXlDRnN5cUVrSWlKTUhnOHJJbWhROHdpZ2dOVDNjSlp0K1hNNis4bURHWDRCK3YrR3JsOEJEUVhwaQpMVVRrcFViWTJETTB6dlhOOWk2em9WQ2JoZUlpMnMzUzdZVmNsV21VY1BoMFpqRFJOVlpQbnhyRzdUeDBvMXlZCnpkdU5vdHhpSHBMNUFOSWRqTXhzUzlSYzdIKys1ekU0eG1GQzJlemlKWCtsWHVlZEVXVT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "ca", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } TASK [fedora.linux_system_roles.certificate : Create return data] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:160 Saturday 14 December 2024 11:33:03 -0500 (0:00:01.376) 0:00:12.680 ***** ok: [managed-node3] => { "ansible_facts": { "certificate_test_certs": { "quadlet_demo": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQTfY97io2RQ2PCiJUq8/sJjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNGRm\nNjNkZWUtMmEzNjQ1MGQtOGYwYTIyNTQtYWJjZmVjMjUwHhcNMjQxMjE0MTYzMzAy\nWhcNMjUxMjE0MTYzMzAxWjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDKJZWUbFEttGr1LGfHPuPpra3GxQ3u0KfF\nOFvEXLpRDKwTp/ma5qcyTUMm7DyZwc/q96lyvfjn8JaUV/SBrvFtI9tKM7RcubQI\nKLYR8LJr414rZWtqjBqzd5cR1JBi0+Id09zx8x6FUlsw+eyXCZPLXgT1Ap8K/sUK\nxNAMl0KhFqXB2mJzR8nfgM5NhvmD+rJcFgebY0e8uzIFAfTon852DkC3mzGEeatt\ncCFVZpCvPkfNBMvIf6vtJkFehMeJ6vOyg1EHYWUVWRwNcFHtf5DsccmLgWZReDIh\n0+KFXuRLb6+Iiww509/eueO3+g28IDJsvVSq4zzkcweoBsVWYjWtAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQU3irQ\n4O9JicdfnsDElgvgOKFqNBUwHwYDVR0jBBgwFoAUaAlZbSqS/YQKTuPAIdmoWZQ9\n2cwwDQYJKoZIhvcNAQELBQADggEBAD4FWxj9I0I1cQ7uaFzmNao1XJs9rww7dHGd\ndHjMpMKNJapY3Q5r1Ew/lRo1MBA+tHVGKLXMK/NjE9MuO8uGbo4zuVq/fuD6mTmh\nbLLFtl+apRm3lpqwqQTuXvTt9EmoWEXhLm50M+OIMyJnrRgm9ZY5pnxCd2zL2rpz\nPWnuI25yCFsyqEkIiJMHg8rImhQ8wiggNT3cJZt+XM6+8mDGX4B+v+Grl8BDQXpi\nLUTkpUbY2DM0zvXN9i6zoVCbheIi2s3S7YVclWmUcPh0ZjDRNVZPnxrG7Tx0o1yY\nzduNotxiHpL5ANIdjMxsS9Rc7H++5zE4xmFC2eziJX+lXuedEWU=\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQTfY97io2RQ2PCiJUq8/sJjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNGRm\nNjNkZWUtMmEzNjQ1MGQtOGYwYTIyNTQtYWJjZmVjMjUwHhcNMjQxMjE0MTYzMzAy\nWhcNMjUxMjE0MTYzMzAxWjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDKJZWUbFEttGr1LGfHPuPpra3GxQ3u0KfF\nOFvEXLpRDKwTp/ma5qcyTUMm7DyZwc/q96lyvfjn8JaUV/SBrvFtI9tKM7RcubQI\nKLYR8LJr414rZWtqjBqzd5cR1JBi0+Id09zx8x6FUlsw+eyXCZPLXgT1Ap8K/sUK\nxNAMl0KhFqXB2mJzR8nfgM5NhvmD+rJcFgebY0e8uzIFAfTon852DkC3mzGEeatt\ncCFVZpCvPkfNBMvIf6vtJkFehMeJ6vOyg1EHYWUVWRwNcFHtf5DsccmLgWZReDIh\n0+KFXuRLb6+Iiww509/eueO3+g28IDJsvVSq4zzkcweoBsVWYjWtAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQU3irQ\n4O9JicdfnsDElgvgOKFqNBUwHwYDVR0jBBgwFoAUaAlZbSqS/YQKTuPAIdmoWZQ9\n2cwwDQYJKoZIhvcNAQELBQADggEBAD4FWxj9I0I1cQ7uaFzmNao1XJs9rww7dHGd\ndHjMpMKNJapY3Q5r1Ew/lRo1MBA+tHVGKLXMK/NjE9MuO8uGbo4zuVq/fuD6mTmh\nbLLFtl+apRm3lpqwqQTuXvTt9EmoWEXhLm50M+OIMyJnrRgm9ZY5pnxCd2zL2rpz\nPWnuI25yCFsyqEkIiJMHg8rImhQ8wiggNT3cJZt+XM6+8mDGX4B+v+Grl8BDQXpi\nLUTkpUbY2DM0zvXN9i6zoVCbheIi2s3S7YVclWmUcPh0ZjDRNVZPnxrG7Tx0o1yY\nzduNotxiHpL5ANIdjMxsS9Rc7H++5zE4xmFC2eziJX+lXuedEWU=\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvgIBADANBgkqhkiG9w0BAQEFAASCBKgwggSkAgEAAoIBAQDKJZWUbFEttGr1\nLGfHPuPpra3GxQ3u0KfFOFvEXLpRDKwTp/ma5qcyTUMm7DyZwc/q96lyvfjn8JaU\nV/SBrvFtI9tKM7RcubQIKLYR8LJr414rZWtqjBqzd5cR1JBi0+Id09zx8x6FUlsw\n+eyXCZPLXgT1Ap8K/sUKxNAMl0KhFqXB2mJzR8nfgM5NhvmD+rJcFgebY0e8uzIF\nAfTon852DkC3mzGEeattcCFVZpCvPkfNBMvIf6vtJkFehMeJ6vOyg1EHYWUVWRwN\ncFHtf5DsccmLgWZReDIh0+KFXuRLb6+Iiww509/eueO3+g28IDJsvVSq4zzkcweo\nBsVWYjWtAgMBAAECggEADZOHn7/Elfa7hjrUv3d0VCHjfWmHnYXq9p5p317A7qeU\nuP87jtE14nV48JjGGyd+oM/AVCH5IQ6x9kWP+NO4iDVzyWB+mQmp0zB0xfCt8vwC\nu5RkqfO96Va7USm/BrwnQAyWFrPW1WpqEHiAkMZGZE6ss43ZA6KHtkjwlh/kjC/8\nI4ARsCbpxdPGahpC47irutUxtapID4wkG6Smfzm873aoewxZDTTuh0Qzpg+t4kr/\nxUmF/QIg5nLlaOXCHB8IwULswKVVvlldA2KURp48OM1AuT1OtVJLaULbobpJ5+mL\nb/+53VA9H5SetWGWB+jsgufnQzyYHEk1wg1Y9rxtwQKBgQDnFabIAvudOxvNGXlB\nc3wmQkPnqZaTnwRQpE1mZTwRI31uhvwBK29/L96uZc/yWNrrdo04OswHqyM11g6T\nG1qdxbHdxHpvbH+5iimy4QmDYPZlronk/bUoAcETcsBZ6ahleCQejgvk2M05U+Uc\nJpF7hym1Ms7XnnO9DGGMWGQ9TQKBgQDf8TLdhAIyW210BbAx6jDu1KQ7hYCkxXEb\nJSlWezu2+JzdJDbXJzmXGgSOddYTkIcgagzzDb0RtteTwqwP+Kk8dkYI/soWr1P4\nHrzMDfkDkWMMPhh0/UZbFWKVkxqQL7ui9k6RBETFUjCU/L1qPdKfhot/4QMhYFNz\nA+x4WtMp4QKBgQDcwNtN8UOLJZ9eZHpoozkzl1QlZ+XX6i4GF591NTqXqM63mq39\nNBBUcVeKROrNOkMQc3VfBRRKMxv03yOXNqdvVx+emobYEM9HYQyKvN4IqsVXx9Al\neJTpvhtDR/knK8baAFaI9eu417/dhx60sMeEOVAHsJzWcKFiNQSvPPWtdQKBgQCK\nYhAelp6M47aFRaaVkbRfQYOjTGIsb93u+k3LJSfHb2XvuZsx5QlPeEJsjiCVwm/d\n9eq+nzTAp44bvuZ4kB50QFoS4O+ciQeIgeYE6nWAYO0l+ZE/aet6CYL8DwURKI34\nq9/s+QhOLvXaezvJVR68eXXmtfMrQs0Oe9Q6o1VtwQKBgFZOL8ZKGFuPpt8e0beK\nauETW4fK3wy0cXyQhrOeczdfddK+K8xi3VrVcQXODT2od4j23NrJ0FbTKUXQ5hG6\nY/dap++omRWklOZ+WPMdiuehEBd0KyyLak8gAHUNcer2BuuHChmL6MIu/j77KPLr\nUbLAnwvoDVZL52JOuTZMEQe7\n-----END PRIVATE KEY-----\n" } } }, "changed": false } TASK [fedora.linux_system_roles.certificate : Stop tracking certificates] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:176 Saturday 14 December 2024 11:33:03 -0500 (0:00:00.063) 0:00:12.743 ***** ok: [managed-node3] => (item={'cert': '/etc/pki/tls/certs/quadlet_demo.crt', 'cert_content': '-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQTfY97io2RQ2PCiJUq8/sJjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNGRm\nNjNkZWUtMmEzNjQ1MGQtOGYwYTIyNTQtYWJjZmVjMjUwHhcNMjQxMjE0MTYzMzAy\nWhcNMjUxMjE0MTYzMzAxWjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDKJZWUbFEttGr1LGfHPuPpra3GxQ3u0KfF\nOFvEXLpRDKwTp/ma5qcyTUMm7DyZwc/q96lyvfjn8JaUV/SBrvFtI9tKM7RcubQI\nKLYR8LJr414rZWtqjBqzd5cR1JBi0+Id09zx8x6FUlsw+eyXCZPLXgT1Ap8K/sUK\nxNAMl0KhFqXB2mJzR8nfgM5NhvmD+rJcFgebY0e8uzIFAfTon852DkC3mzGEeatt\ncCFVZpCvPkfNBMvIf6vtJkFehMeJ6vOyg1EHYWUVWRwNcFHtf5DsccmLgWZReDIh\n0+KFXuRLb6+Iiww509/eueO3+g28IDJsvVSq4zzkcweoBsVWYjWtAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQU3irQ\n4O9JicdfnsDElgvgOKFqNBUwHwYDVR0jBBgwFoAUaAlZbSqS/YQKTuPAIdmoWZQ9\n2cwwDQYJKoZIhvcNAQELBQADggEBAD4FWxj9I0I1cQ7uaFzmNao1XJs9rww7dHGd\ndHjMpMKNJapY3Q5r1Ew/lRo1MBA+tHVGKLXMK/NjE9MuO8uGbo4zuVq/fuD6mTmh\nbLLFtl+apRm3lpqwqQTuXvTt9EmoWEXhLm50M+OIMyJnrRgm9ZY5pnxCd2zL2rpz\nPWnuI25yCFsyqEkIiJMHg8rImhQ8wiggNT3cJZt+XM6+8mDGX4B+v+Grl8BDQXpi\nLUTkpUbY2DM0zvXN9i6zoVCbheIi2s3S7YVclWmUcPh0ZjDRNVZPnxrG7Tx0o1yY\nzduNotxiHpL5ANIdjMxsS9Rc7H++5zE4xmFC2eziJX+lXuedEWU=\n-----END CERTIFICATE-----\n', 'key': '/etc/pki/tls/private/quadlet_demo.key', 'key_content': '-----BEGIN PRIVATE KEY-----\nMIIEvgIBADANBgkqhkiG9w0BAQEFAASCBKgwggSkAgEAAoIBAQDKJZWUbFEttGr1\nLGfHPuPpra3GxQ3u0KfFOFvEXLpRDKwTp/ma5qcyTUMm7DyZwc/q96lyvfjn8JaU\nV/SBrvFtI9tKM7RcubQIKLYR8LJr414rZWtqjBqzd5cR1JBi0+Id09zx8x6FUlsw\n+eyXCZPLXgT1Ap8K/sUKxNAMl0KhFqXB2mJzR8nfgM5NhvmD+rJcFgebY0e8uzIF\nAfTon852DkC3mzGEeattcCFVZpCvPkfNBMvIf6vtJkFehMeJ6vOyg1EHYWUVWRwN\ncFHtf5DsccmLgWZReDIh0+KFXuRLb6+Iiww509/eueO3+g28IDJsvVSq4zzkcweo\nBsVWYjWtAgMBAAECggEADZOHn7/Elfa7hjrUv3d0VCHjfWmHnYXq9p5p317A7qeU\nuP87jtE14nV48JjGGyd+oM/AVCH5IQ6x9kWP+NO4iDVzyWB+mQmp0zB0xfCt8vwC\nu5RkqfO96Va7USm/BrwnQAyWFrPW1WpqEHiAkMZGZE6ss43ZA6KHtkjwlh/kjC/8\nI4ARsCbpxdPGahpC47irutUxtapID4wkG6Smfzm873aoewxZDTTuh0Qzpg+t4kr/\nxUmF/QIg5nLlaOXCHB8IwULswKVVvlldA2KURp48OM1AuT1OtVJLaULbobpJ5+mL\nb/+53VA9H5SetWGWB+jsgufnQzyYHEk1wg1Y9rxtwQKBgQDnFabIAvudOxvNGXlB\nc3wmQkPnqZaTnwRQpE1mZTwRI31uhvwBK29/L96uZc/yWNrrdo04OswHqyM11g6T\nG1qdxbHdxHpvbH+5iimy4QmDYPZlronk/bUoAcETcsBZ6ahleCQejgvk2M05U+Uc\nJpF7hym1Ms7XnnO9DGGMWGQ9TQKBgQDf8TLdhAIyW210BbAx6jDu1KQ7hYCkxXEb\nJSlWezu2+JzdJDbXJzmXGgSOddYTkIcgagzzDb0RtteTwqwP+Kk8dkYI/soWr1P4\nHrzMDfkDkWMMPhh0/UZbFWKVkxqQL7ui9k6RBETFUjCU/L1qPdKfhot/4QMhYFNz\nA+x4WtMp4QKBgQDcwNtN8UOLJZ9eZHpoozkzl1QlZ+XX6i4GF591NTqXqM63mq39\nNBBUcVeKROrNOkMQc3VfBRRKMxv03yOXNqdvVx+emobYEM9HYQyKvN4IqsVXx9Al\neJTpvhtDR/knK8baAFaI9eu417/dhx60sMeEOVAHsJzWcKFiNQSvPPWtdQKBgQCK\nYhAelp6M47aFRaaVkbRfQYOjTGIsb93u+k3LJSfHb2XvuZsx5QlPeEJsjiCVwm/d\n9eq+nzTAp44bvuZ4kB50QFoS4O+ciQeIgeYE6nWAYO0l+ZE/aet6CYL8DwURKI34\nq9/s+QhOLvXaezvJVR68eXXmtfMrQs0Oe9Q6o1VtwQKBgFZOL8ZKGFuPpt8e0beK\nauETW4fK3wy0cXyQhrOeczdfddK+K8xi3VrVcQXODT2od4j23NrJ0FbTKUXQ5hG6\nY/dap++omRWklOZ+WPMdiuehEBd0KyyLak8gAHUNcer2BuuHChmL6MIu/j77KPLr\nUbLAnwvoDVZL52JOuTZMEQe7\n-----END PRIVATE KEY-----\n', 'ca': '/etc/pki/tls/certs/quadlet_demo.crt', 'ca_content': '-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQTfY97io2RQ2PCiJUq8/sJjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNGRm\nNjNkZWUtMmEzNjQ1MGQtOGYwYTIyNTQtYWJjZmVjMjUwHhcNMjQxMjE0MTYzMzAy\nWhcNMjUxMjE0MTYzMzAxWjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDKJZWUbFEttGr1LGfHPuPpra3GxQ3u0KfF\nOFvEXLpRDKwTp/ma5qcyTUMm7DyZwc/q96lyvfjn8JaUV/SBrvFtI9tKM7RcubQI\nKLYR8LJr414rZWtqjBqzd5cR1JBi0+Id09zx8x6FUlsw+eyXCZPLXgT1Ap8K/sUK\nxNAMl0KhFqXB2mJzR8nfgM5NhvmD+rJcFgebY0e8uzIFAfTon852DkC3mzGEeatt\ncCFVZpCvPkfNBMvIf6vtJkFehMeJ6vOyg1EHYWUVWRwNcFHtf5DsccmLgWZReDIh\n0+KFXuRLb6+Iiww509/eueO3+g28IDJsvVSq4zzkcweoBsVWYjWtAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQU3irQ\n4O9JicdfnsDElgvgOKFqNBUwHwYDVR0jBBgwFoAUaAlZbSqS/YQKTuPAIdmoWZQ9\n2cwwDQYJKoZIhvcNAQELBQADggEBAD4FWxj9I0I1cQ7uaFzmNao1XJs9rww7dHGd\ndHjMpMKNJapY3Q5r1Ew/lRo1MBA+tHVGKLXMK/NjE9MuO8uGbo4zuVq/fuD6mTmh\nbLLFtl+apRm3lpqwqQTuXvTt9EmoWEXhLm50M+OIMyJnrRgm9ZY5pnxCd2zL2rpz\nPWnuI25yCFsyqEkIiJMHg8rImhQ8wiggNT3cJZt+XM6+8mDGX4B+v+Grl8BDQXpi\nLUTkpUbY2DM0zvXN9i6zoVCbheIi2s3S7YVclWmUcPh0ZjDRNVZPnxrG7Tx0o1yY\nzduNotxiHpL5ANIdjMxsS9Rc7H++5zE4xmFC2eziJX+lXuedEWU=\n-----END CERTIFICATE-----\n'}) => { "ansible_loop_var": "item", "changed": false, "cmd": [ "getcert", "stop-tracking", "-f", "/etc/pki/tls/certs/quadlet_demo.crt" ], "delta": "0:00:00.026207", "end": "2024-12-14 11:33:04.279294", "item": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQTfY97io2RQ2PCiJUq8/sJjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNGRm\nNjNkZWUtMmEzNjQ1MGQtOGYwYTIyNTQtYWJjZmVjMjUwHhcNMjQxMjE0MTYzMzAy\nWhcNMjUxMjE0MTYzMzAxWjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDKJZWUbFEttGr1LGfHPuPpra3GxQ3u0KfF\nOFvEXLpRDKwTp/ma5qcyTUMm7DyZwc/q96lyvfjn8JaUV/SBrvFtI9tKM7RcubQI\nKLYR8LJr414rZWtqjBqzd5cR1JBi0+Id09zx8x6FUlsw+eyXCZPLXgT1Ap8K/sUK\nxNAMl0KhFqXB2mJzR8nfgM5NhvmD+rJcFgebY0e8uzIFAfTon852DkC3mzGEeatt\ncCFVZpCvPkfNBMvIf6vtJkFehMeJ6vOyg1EHYWUVWRwNcFHtf5DsccmLgWZReDIh\n0+KFXuRLb6+Iiww509/eueO3+g28IDJsvVSq4zzkcweoBsVWYjWtAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQU3irQ\n4O9JicdfnsDElgvgOKFqNBUwHwYDVR0jBBgwFoAUaAlZbSqS/YQKTuPAIdmoWZQ9\n2cwwDQYJKoZIhvcNAQELBQADggEBAD4FWxj9I0I1cQ7uaFzmNao1XJs9rww7dHGd\ndHjMpMKNJapY3Q5r1Ew/lRo1MBA+tHVGKLXMK/NjE9MuO8uGbo4zuVq/fuD6mTmh\nbLLFtl+apRm3lpqwqQTuXvTt9EmoWEXhLm50M+OIMyJnrRgm9ZY5pnxCd2zL2rpz\nPWnuI25yCFsyqEkIiJMHg8rImhQ8wiggNT3cJZt+XM6+8mDGX4B+v+Grl8BDQXpi\nLUTkpUbY2DM0zvXN9i6zoVCbheIi2s3S7YVclWmUcPh0ZjDRNVZPnxrG7Tx0o1yY\nzduNotxiHpL5ANIdjMxsS9Rc7H++5zE4xmFC2eziJX+lXuedEWU=\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgjCCAmqgAwIBAgIQTfY97io2RQ2PCiJUq8/sJjANBgkqhkiG9w0BAQsFADBQ\nMSAwHgYDVQQDDBdMb2NhbCBTaWduaW5nIEF1dGhvcml0eTEsMCoGA1UEAwwjNGRm\nNjNkZWUtMmEzNjQ1MGQtOGYwYTIyNTQtYWJjZmVjMjUwHhcNMjQxMjE0MTYzMzAy\nWhcNMjUxMjE0MTYzMzAxWjAUMRIwEAYDVQQDEwlsb2NhbGhvc3QwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDKJZWUbFEttGr1LGfHPuPpra3GxQ3u0KfF\nOFvEXLpRDKwTp/ma5qcyTUMm7DyZwc/q96lyvfjn8JaUV/SBrvFtI9tKM7RcubQI\nKLYR8LJr414rZWtqjBqzd5cR1JBi0+Id09zx8x6FUlsw+eyXCZPLXgT1Ap8K/sUK\nxNAMl0KhFqXB2mJzR8nfgM5NhvmD+rJcFgebY0e8uzIFAfTon852DkC3mzGEeatt\ncCFVZpCvPkfNBMvIf6vtJkFehMeJ6vOyg1EHYWUVWRwNcFHtf5DsccmLgWZReDIh\n0+KFXuRLb6+Iiww509/eueO3+g28IDJsvVSq4zzkcweoBsVWYjWtAgMBAAGjgZMw\ngZAwCwYDVR0PBAQDAgWgMBQGA1UdEQQNMAuCCWxvY2FsaG9zdDAdBgNVHSUEFjAU\nBggrBgEFBQcDAQYIKwYBBQUHAwIwDAYDVR0TAQH/BAIwADAdBgNVHQ4EFgQU3irQ\n4O9JicdfnsDElgvgOKFqNBUwHwYDVR0jBBgwFoAUaAlZbSqS/YQKTuPAIdmoWZQ9\n2cwwDQYJKoZIhvcNAQELBQADggEBAD4FWxj9I0I1cQ7uaFzmNao1XJs9rww7dHGd\ndHjMpMKNJapY3Q5r1Ew/lRo1MBA+tHVGKLXMK/NjE9MuO8uGbo4zuVq/fuD6mTmh\nbLLFtl+apRm3lpqwqQTuXvTt9EmoWEXhLm50M+OIMyJnrRgm9ZY5pnxCd2zL2rpz\nPWnuI25yCFsyqEkIiJMHg8rImhQ8wiggNT3cJZt+XM6+8mDGX4B+v+Grl8BDQXpi\nLUTkpUbY2DM0zvXN9i6zoVCbheIi2s3S7YVclWmUcPh0ZjDRNVZPnxrG7Tx0o1yY\nzduNotxiHpL5ANIdjMxsS9Rc7H++5zE4xmFC2eziJX+lXuedEWU=\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvgIBADANBgkqhkiG9w0BAQEFAASCBKgwggSkAgEAAoIBAQDKJZWUbFEttGr1\nLGfHPuPpra3GxQ3u0KfFOFvEXLpRDKwTp/ma5qcyTUMm7DyZwc/q96lyvfjn8JaU\nV/SBrvFtI9tKM7RcubQIKLYR8LJr414rZWtqjBqzd5cR1JBi0+Id09zx8x6FUlsw\n+eyXCZPLXgT1Ap8K/sUKxNAMl0KhFqXB2mJzR8nfgM5NhvmD+rJcFgebY0e8uzIF\nAfTon852DkC3mzGEeattcCFVZpCvPkfNBMvIf6vtJkFehMeJ6vOyg1EHYWUVWRwN\ncFHtf5DsccmLgWZReDIh0+KFXuRLb6+Iiww509/eueO3+g28IDJsvVSq4zzkcweo\nBsVWYjWtAgMBAAECggEADZOHn7/Elfa7hjrUv3d0VCHjfWmHnYXq9p5p317A7qeU\nuP87jtE14nV48JjGGyd+oM/AVCH5IQ6x9kWP+NO4iDVzyWB+mQmp0zB0xfCt8vwC\nu5RkqfO96Va7USm/BrwnQAyWFrPW1WpqEHiAkMZGZE6ss43ZA6KHtkjwlh/kjC/8\nI4ARsCbpxdPGahpC47irutUxtapID4wkG6Smfzm873aoewxZDTTuh0Qzpg+t4kr/\nxUmF/QIg5nLlaOXCHB8IwULswKVVvlldA2KURp48OM1AuT1OtVJLaULbobpJ5+mL\nb/+53VA9H5SetWGWB+jsgufnQzyYHEk1wg1Y9rxtwQKBgQDnFabIAvudOxvNGXlB\nc3wmQkPnqZaTnwRQpE1mZTwRI31uhvwBK29/L96uZc/yWNrrdo04OswHqyM11g6T\nG1qdxbHdxHpvbH+5iimy4QmDYPZlronk/bUoAcETcsBZ6ahleCQejgvk2M05U+Uc\nJpF7hym1Ms7XnnO9DGGMWGQ9TQKBgQDf8TLdhAIyW210BbAx6jDu1KQ7hYCkxXEb\nJSlWezu2+JzdJDbXJzmXGgSOddYTkIcgagzzDb0RtteTwqwP+Kk8dkYI/soWr1P4\nHrzMDfkDkWMMPhh0/UZbFWKVkxqQL7ui9k6RBETFUjCU/L1qPdKfhot/4QMhYFNz\nA+x4WtMp4QKBgQDcwNtN8UOLJZ9eZHpoozkzl1QlZ+XX6i4GF591NTqXqM63mq39\nNBBUcVeKROrNOkMQc3VfBRRKMxv03yOXNqdvVx+emobYEM9HYQyKvN4IqsVXx9Al\neJTpvhtDR/knK8baAFaI9eu417/dhx60sMeEOVAHsJzWcKFiNQSvPPWtdQKBgQCK\nYhAelp6M47aFRaaVkbRfQYOjTGIsb93u+k3LJSfHb2XvuZsx5QlPeEJsjiCVwm/d\n9eq+nzTAp44bvuZ4kB50QFoS4O+ciQeIgeYE6nWAYO0l+ZE/aet6CYL8DwURKI34\nq9/s+QhOLvXaezvJVR68eXXmtfMrQs0Oe9Q6o1VtwQKBgFZOL8ZKGFuPpt8e0beK\nauETW4fK3wy0cXyQhrOeczdfddK+K8xi3VrVcQXODT2od4j23NrJ0FbTKUXQ5hG6\nY/dap++omRWklOZ+WPMdiuehEBd0KyyLak8gAHUNcer2BuuHChmL6MIu/j77KPLr\nUbLAnwvoDVZL52JOuTZMEQe7\n-----END PRIVATE KEY-----\n" }, "rc": 0, "start": "2024-12-14 11:33:04.253087" } STDOUT: Request "20241214163301" removed. TASK [fedora.linux_system_roles.certificate : Remove files] ******************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181 Saturday 14 December 2024 11:33:04 -0500 (0:00:00.638) 0:00:13.382 ***** changed: [managed-node3] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } changed: [managed-node3] => (item=/etc/pki/tls/private/quadlet_demo.key) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/private/quadlet_demo.key", "path": "/etc/pki/tls/private/quadlet_demo.key", "state": "absent" } ok: [managed-node3] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } TASK [Run the role] ************************************************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:62 Saturday 14 December 2024 11:33:05 -0500 (0:00:01.303) 0:00:14.685 ***** included: fedora.linux_system_roles.podman for managed-node3 TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Saturday 14 December 2024 11:33:05 -0500 (0:00:00.190) 0:00:14.876 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Saturday 14 December 2024 11:33:05 -0500 (0:00:00.049) 0:00:14.925 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Saturday 14 December 2024 11:33:05 -0500 (0:00:00.074) 0:00:14.999 ***** ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Saturday 14 December 2024 11:33:06 -0500 (0:00:00.391) 0:00:15.390 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Saturday 14 December 2024 11:33:06 -0500 (0:00:00.024) 0:00:15.415 ***** ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Saturday 14 December 2024 11:33:06 -0500 (0:00:00.395) 0:00:15.810 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_is_transactional": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Saturday 14 December 2024 11:33:06 -0500 (0:00:00.054) 0:00:15.864 ***** ok: [managed-node3] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed-node3] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node3] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } ok: [managed-node3] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Saturday 14 December 2024 11:33:06 -0500 (0:00:00.076) 0:00:15.940 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Saturday 14 December 2024 11:33:08 -0500 (0:00:01.371) 0:00:17.311 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Saturday 14 December 2024 11:33:08 -0500 (0:00:00.076) 0:00:17.388 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages))", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Saturday 14 December 2024 11:33:08 -0500 (0:00:00.086) 0:00:17.474 ***** skipping: [managed-node3] => { "false_condition": "__podman_is_transactional | d(false)" } TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Saturday 14 December 2024 11:33:08 -0500 (0:00:00.075) 0:00:17.550 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Saturday 14 December 2024 11:33:08 -0500 (0:00:00.078) 0:00:17.628 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Saturday 14 December 2024 11:33:08 -0500 (0:00:00.063) 0:00:17.692 ***** ok: [managed-node3] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.028811", "end": "2024-12-14 11:33:09.025084", "rc": 0, "start": "2024-12-14 11:33:08.996273" } STDOUT: podman version 5.3.1 STDERR: time="2024-12-14T11:33:09-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Saturday 14 December 2024 11:33:09 -0500 (0:00:00.454) 0:00:18.148 ***** ok: [managed-node3] => { "ansible_facts": { "podman_version": "5.3.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Saturday 14 December 2024 11:33:09 -0500 (0:00:00.076) 0:00:18.224 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Saturday 14 December 2024 11:33:09 -0500 (0:00:00.111) 0:00:18.336 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Saturday 14 December 2024 11:33:09 -0500 (0:00:00.178) 0:00:18.514 ***** META: end_host conditional evaluated to False, continuing execution for managed-node3 skipping: [managed-node3] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node3" } MSG: end_host conditional evaluated to false, continuing execution for managed-node3 TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Saturday 14 December 2024 11:33:09 -0500 (0:00:00.135) 0:00:18.650 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__has_type_pod or __has_pod_file_ext or __has_pod_file_src_ext or __has_pod_template_src_ext or __has_pod_template_src_ext_j2", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:96 Saturday 14 December 2024 11:33:09 -0500 (0:00:00.043) 0:00:18.694 ***** META: end_host conditional evaluated to False, continuing execution for managed-node3 skipping: [managed-node3] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node3" } MSG: end_host conditional evaluated to false, continuing execution for managed-node3 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:109 Saturday 14 December 2024 11:33:09 -0500 (0:00:00.050) 0:00:18.745 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 14 December 2024 11:33:09 -0500 (0:00:00.108) 0:00:18.853 ***** ok: [managed-node3] => { "ansible_facts": { "getent_passwd": { "root": [ "x", "0", "0", "Super User", "/root", "/bin/bash" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 14 December 2024 11:33:10 -0500 (0:00:00.614) 0:00:19.468 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 14 December 2024 11:33:10 -0500 (0:00:00.086) 0:00:19.555 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 14 December 2024 11:33:10 -0500 (0:00:00.079) 0:00:19.635 ***** ok: [managed-node3] => { "changed": false, "stat": { "atime": 1734193882.6705027, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1734193862.9755826, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 8859182, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "2878164177", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 14 December 2024 11:33:11 -0500 (0:00:00.488) 0:00:20.123 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:44 Saturday 14 December 2024 11:33:11 -0500 (0:00:00.103) 0:00:20.227 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Saturday 14 December 2024 11:33:11 -0500 (0:00:00.064) 0:00:20.292 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:62 Saturday 14 December 2024 11:33:11 -0500 (0:00:00.065) 0:00:20.357 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:67 Saturday 14 December 2024 11:33:11 -0500 (0:00:00.067) 0:00:20.424 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:72 Saturday 14 December 2024 11:33:11 -0500 (0:00:00.069) 0:00:20.493 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:82 Saturday 14 December 2024 11:33:11 -0500 (0:00:00.044) 0:00:20.538 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:89 Saturday 14 December 2024 11:33:11 -0500 (0:00:00.065) 0:00:20.603 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:115 Saturday 14 December 2024 11:33:11 -0500 (0:00:00.056) 0:00:20.660 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:124 Saturday 14 December 2024 11:33:11 -0500 (0:00:00.115) 0:00:20.776 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Saturday 14 December 2024 11:33:11 -0500 (0:00:00.110) 0:00:20.886 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Saturday 14 December 2024 11:33:11 -0500 (0:00:00.066) 0:00:20.952 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:127 Saturday 14 December 2024 11:33:12 -0500 (0:00:00.098) 0:00:21.051 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Saturday 14 December 2024 11:33:12 -0500 (0:00:00.092) 0:00:21.144 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Saturday 14 December 2024 11:33:12 -0500 (0:00:00.050) 0:00:21.194 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:130 Saturday 14 December 2024 11:33:12 -0500 (0:00:00.051) 0:00:21.245 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Saturday 14 December 2024 11:33:12 -0500 (0:00:00.100) 0:00:21.346 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Saturday 14 December 2024 11:33:12 -0500 (0:00:00.051) 0:00:21.397 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:133 Saturday 14 December 2024 11:33:12 -0500 (0:00:00.053) 0:00:21.451 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Saturday 14 December 2024 11:33:12 -0500 (0:00:00.102) 0:00:21.554 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Saturday 14 December 2024 11:33:12 -0500 (0:00:00.074) 0:00:21.629 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Saturday 14 December 2024 11:33:12 -0500 (0:00:00.081) 0:00:21.710 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Saturday 14 December 2024 11:33:12 -0500 (0:00:00.106) 0:00:21.816 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:139 Saturday 14 December 2024 11:33:12 -0500 (0:00:00.053) 0:00:21.870 ***** included: fedora.linux_system_roles.firewall for managed-node3 TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Saturday 14 December 2024 11:33:13 -0500 (0:00:00.182) 0:00:22.053 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed-node3 TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Saturday 14 December 2024 11:33:13 -0500 (0:00:00.091) 0:00:22.144 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Saturday 14 December 2024 11:33:13 -0500 (0:00:00.061) 0:00:22.206 ***** ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Saturday 14 December 2024 11:33:13 -0500 (0:00:00.530) 0:00:22.737 ***** ok: [managed-node3] => { "ansible_facts": { "__firewall_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22 Saturday 14 December 2024 11:33:13 -0500 (0:00:00.105) 0:00:22.843 ***** ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27 Saturday 14 December 2024 11:33:14 -0500 (0:00:00.487) 0:00:23.330 ***** ok: [managed-node3] => { "ansible_facts": { "__firewall_is_transactional": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 Saturday 14 December 2024 11:33:14 -0500 (0:00:00.083) 0:00:23.414 ***** ok: [managed-node3] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: firewalld TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43 Saturday 14 December 2024 11:33:15 -0500 (0:00:00.894) 0:00:24.308 ***** skipping: [managed-node3] => { "false_condition": "__firewall_is_transactional | d(false)" } TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48 Saturday 14 December 2024 11:33:15 -0500 (0:00:00.101) 0:00:24.410 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53 Saturday 14 December 2024 11:33:15 -0500 (0:00:00.057) 0:00:24.467 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Saturday 14 December 2024 11:33:15 -0500 (0:00:00.088) 0:00:24.556 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Saturday 14 December 2024 11:33:15 -0500 (0:00:00.060) 0:00:24.616 ***** skipping: [managed-node3] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Saturday 14 December 2024 11:33:15 -0500 (0:00:00.062) 0:00:24.679 ***** ok: [managed-node3] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "dbus.socket basic.target polkit.service system.slice dbus-broker.service sysinit.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target ipset.service iptables.service ebtables.service ip6tables.service", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698237440", "EffectiveMemoryMax": "3698237440", "EffectiveTasksMax": "22361", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3231113216", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "[not set]", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "[not set]", "MemorySwapMax": "infinity", "MemorySwapPeak": "[not set]", "MemoryZSwapCurrent": "[not set]", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22361", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Saturday 14 December 2024 11:33:16 -0500 (0:00:00.611) 0:00:25.290 ***** changed: [managed-node3] => { "changed": true, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "basic.target system.slice sysinit.target dbus.socket dbus-broker.service polkit.service", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target network-pre.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "ebtables.service iptables.service ipset.service shutdown.target ip6tables.service", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveMemoryHigh": "3698237440", "EffectiveMemoryMax": "3698237440", "EffectiveTasksMax": "22361", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3230429184", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "[not set]", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "[not set]", "MemorySwapMax": "infinity", "MemorySwapPeak": "[not set]", "MemoryZSwapCurrent": "[not set]", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22361", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "disabled", "UtmpMode": "init", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Saturday 14 December 2024 11:33:17 -0500 (0:00:01.102) 0:00:26.393 ***** ok: [managed-node3] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3.12", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Saturday 14 December 2024 11:33:17 -0500 (0:00:00.048) 0:00:26.442 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Saturday 14 December 2024 11:33:17 -0500 (0:00:00.028) 0:00:26.471 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Saturday 14 December 2024 11:33:17 -0500 (0:00:00.033) 0:00:26.504 ***** changed: [managed-node3] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "8000/tcp", "state": "enabled" } } changed: [managed-node3] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Saturday 14 December 2024 11:33:18 -0500 (0:00:01.295) 0:00:27.799 ***** skipping: [managed-node3] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed-node3] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Saturday 14 December 2024 11:33:18 -0500 (0:00:00.046) 0:00:27.845 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "firewall | length == 1", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Saturday 14 December 2024 11:33:18 -0500 (0:00:00.033) 0:00:27.879 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Saturday 14 December 2024 11:33:18 -0500 (0:00:00.030) 0:00:27.910 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Saturday 14 December 2024 11:33:18 -0500 (0:00:00.031) 0:00:27.941 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Saturday 14 December 2024 11:33:18 -0500 (0:00:00.028) 0:00:27.970 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Saturday 14 December 2024 11:33:18 -0500 (0:00:00.030) 0:00:28.001 ***** skipping: [managed-node3] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:146 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.080) 0:00:28.081 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:153 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.036) 0:00:28.117 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:157 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.043) 0:00:28.161 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:166 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.044) 0:00:28.205 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:175 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.031) 0:00:28.236 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed-node3 => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.144) 0:00:28.381 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.034) 0:00:28.415 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.045) 0:00:28.461 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.082) 0:00:28.543 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.045) 0:00:28.589 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.074) 0:00:28.664 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.056) 0:00:28.721 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Saturday 14 December 2024 11:33:19 -0500 (0:00:00.050) 0:00:28.772 ***** [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Saturday 14 December 2024 11:33:20 -0500 (0:00:00.803) 0:00:29.575 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Saturday 14 December 2024 11:33:20 -0500 (0:00:00.065) 0:00:29.641 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Saturday 14 December 2024 11:33:20 -0500 (0:00:00.061) 0:00:29.702 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:33:20 -0500 (0:00:00.080) 0:00:29.783 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:33:20 -0500 (0:00:00.162) 0:00:29.946 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:33:20 -0500 (0:00:00.058) 0:00:30.004 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Saturday 14 December 2024 11:33:21 -0500 (0:00:00.060) 0:00:30.065 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Saturday 14 December 2024 11:33:21 -0500 (0:00:00.055) 0:00:30.121 ***** changed: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Saturday 14 December 2024 11:33:21 -0500 (0:00:00.539) 0:00:30.660 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Saturday 14 December 2024 11:33:21 -0500 (0:00:00.039) 0:00:30.700 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Saturday 14 December 2024 11:33:21 -0500 (0:00:00.058) 0:00:30.759 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:33:21 -0500 (0:00:00.071) 0:00:30.830 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:33:21 -0500 (0:00:00.050) 0:00:30.881 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:33:21 -0500 (0:00:00.048) 0:00:30.930 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Saturday 14 December 2024 11:33:21 -0500 (0:00:00.054) 0:00:30.984 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Saturday 14 December 2024 11:33:21 -0500 (0:00:00.045) 0:00:31.029 ***** changed: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:182 Saturday 14 December 2024 11:33:22 -0500 (0:00:00.559) 0:00:31.589 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:189 Saturday 14 December 2024 11:33:22 -0500 (0:00:00.056) 0:00:31.645 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node3 => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 14 December 2024 11:33:23 -0500 (0:00:00.412) 0:00:32.057 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 14 December 2024 11:33:23 -0500 (0:00:00.117) 0:00:32.175 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 14 December 2024 11:33:23 -0500 (0:00:00.118) 0:00:32.294 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 14 December 2024 11:33:23 -0500 (0:00:00.097) 0:00:32.391 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 14 December 2024 11:33:23 -0500 (0:00:00.253) 0:00:32.645 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 14 December 2024 11:33:23 -0500 (0:00:00.120) 0:00:32.766 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 14 December 2024 11:33:23 -0500 (0:00:00.066) 0:00:32.832 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 14 December 2024 11:33:23 -0500 (0:00:00.065) 0:00:32.897 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 14 December 2024 11:33:23 -0500 (0:00:00.058) 0:00:32.956 ***** ok: [managed-node3] => { "changed": false, "stat": { "atime": 1734193882.6705027, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1734193862.9755826, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 8859182, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "2878164177", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 14 December 2024 11:33:24 -0500 (0:00:00.401) 0:00:33.358 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:44 Saturday 14 December 2024 11:33:24 -0500 (0:00:00.028) 0:00:33.386 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Saturday 14 December 2024 11:33:24 -0500 (0:00:00.030) 0:00:33.416 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:62 Saturday 14 December 2024 11:33:24 -0500 (0:00:00.028) 0:00:33.444 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:67 Saturday 14 December 2024 11:33:24 -0500 (0:00:00.032) 0:00:33.477 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:72 Saturday 14 December 2024 11:33:24 -0500 (0:00:00.038) 0:00:33.515 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:82 Saturday 14 December 2024 11:33:24 -0500 (0:00:00.044) 0:00:33.560 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:89 Saturday 14 December 2024 11:33:24 -0500 (0:00:00.081) 0:00:33.641 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 14 December 2024 11:33:24 -0500 (0:00:00.076) 0:00:33.718 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-network.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 14 December 2024 11:33:24 -0500 (0:00:00.131) 0:00:33.850 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 14 December 2024 11:33:24 -0500 (0:00:00.062) 0:00:33.912 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 14 December 2024 11:33:24 -0500 (0:00:00.053) 0:00:33.966 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:105 Saturday 14 December 2024 11:33:25 -0500 (0:00:00.125) 0:00:34.092 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:112 Saturday 14 December 2024 11:33:25 -0500 (0:00:00.082) 0:00:34.174 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:116 Saturday 14 December 2024 11:33:25 -0500 (0:00:00.046) 0:00:34.221 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Saturday 14 December 2024 11:33:25 -0500 (0:00:00.120) 0:00:34.341 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:33:25 -0500 (0:00:00.050) 0:00:34.392 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:33:25 -0500 (0:00:00.030) 0:00:34.422 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:33:25 -0500 (0:00:00.030) 0:00:34.453 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Saturday 14 December 2024 11:33:25 -0500 (0:00:00.040) 0:00:34.493 ***** skipping: [managed-node3] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Saturday 14 December 2024 11:33:25 -0500 (0:00:00.042) 0:00:34.536 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Saturday 14 December 2024 11:33:25 -0500 (0:00:00.034) 0:00:34.570 ***** ok: [managed-node3] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Saturday 14 December 2024 11:33:25 -0500 (0:00:00.438) 0:00:35.009 ***** changed: [managed-node3] => { "changed": true, "checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0", "dest": "/etc/containers/systemd/quadlet-demo.network", "gid": 0, "group": "root", "md5sum": "061f3cf318cbd8ab5794bb1173831fb8", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 74, "src": "/root/.ansible/tmp/ansible-tmp-1734194006.0269196-14057-133971458264709/.source.network", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Saturday 14 December 2024 11:33:26 -0500 (0:00:00.863) 0:00:35.872 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Saturday 14 December 2024 11:33:26 -0500 (0:00:00.052) 0:00:35.924 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Saturday 14 December 2024 11:33:26 -0500 (0:00:00.052) 0:00:35.976 ***** ok: [managed-node3] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Saturday 14 December 2024 11:33:27 -0500 (0:00:00.767) 0:00:36.744 ***** changed: [managed-node3] => { "changed": true, "name": "quadlet-demo-network.service", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "-.mount system.slice systemd-journald.socket sysinit.target network-online.target basic.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "quadlet-demo-network.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveMemoryHigh": "3698237440", "EffectiveMemoryMax": "3698237440", "EffectiveTasksMax": "22361", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet 192.168.30.0/24 --gateway 192.168.30.1 --label app=wordpress systemd-quadlet-demo ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet 192.168.30.0/24 --gateway 192.168.30.1 --label app=wordpress systemd-quadlet-demo ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-network.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-network.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3195224064", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "[not set]", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "[not set]", "MemorySwapMax": "infinity", "MemorySwapPeak": "[not set]", "MemoryZSwapCurrent": "[not set]", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-network.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "system.slice -.mount sysinit.target", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo.network", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-network", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22361", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "Wants": "network-online.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Saturday 14 December 2024 11:33:28 -0500 (0:00:00.633) 0:00:37.377 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_service_started is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 14 December 2024 11:33:28 -0500 (0:00:00.030) 0:00:37.408 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 14 December 2024 11:33:28 -0500 (0:00:00.042) 0:00:37.451 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 14 December 2024 11:33:28 -0500 (0:00:00.038) 0:00:37.490 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 14 December 2024 11:33:28 -0500 (0:00:00.034) 0:00:37.524 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 14 December 2024 11:33:28 -0500 (0:00:00.075) 0:00:37.600 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 14 December 2024 11:33:28 -0500 (0:00:00.096) 0:00:37.696 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 14 December 2024 11:33:28 -0500 (0:00:00.124) 0:00:37.820 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 14 December 2024 11:33:28 -0500 (0:00:00.055) 0:00:37.875 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 14 December 2024 11:33:28 -0500 (0:00:00.069) 0:00:37.945 ***** ok: [managed-node3] => { "changed": false, "stat": { "atime": 1734193882.6705027, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1734193862.9755826, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 8859182, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "2878164177", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.419) 0:00:38.365 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:44 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.043) 0:00:38.409 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.035) 0:00:38.444 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:62 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.032) 0:00:38.476 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:67 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.035) 0:00:38.512 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:72 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.040) 0:00:38.552 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:82 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.034) 0:00:38.586 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:89 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.029) 0:00:38.616 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.028) 0:00:38.645 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.054) 0:00:38.699 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.033) 0:00:38.732 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.041) 0:00:38.774 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:105 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.073) 0:00:38.848 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:112 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.039) 0:00:38.887 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:116 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.029) 0:00:38.916 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Saturday 14 December 2024 11:33:29 -0500 (0:00:00.064) 0:00:38.981 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:33:30 -0500 (0:00:00.088) 0:00:39.070 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:33:30 -0500 (0:00:00.029) 0:00:39.100 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:33:30 -0500 (0:00:00.036) 0:00:39.136 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Saturday 14 December 2024 11:33:30 -0500 (0:00:00.037) 0:00:39.173 ***** skipping: [managed-node3] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Saturday 14 December 2024 11:33:30 -0500 (0:00:00.046) 0:00:39.220 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Saturday 14 December 2024 11:33:30 -0500 (0:00:00.034) 0:00:39.255 ***** ok: [managed-node3] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 34, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Saturday 14 December 2024 11:33:30 -0500 (0:00:00.404) 0:00:39.660 ***** changed: [managed-node3] => { "changed": true, "checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a", "dest": "/etc/containers/systemd/quadlet-demo-mysql.volume", "gid": 0, "group": "root", "md5sum": "5ddd03a022aeb4502d9bc8ce436b4233", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 9, "src": "/root/.ansible/tmp/ansible-tmp-1734194010.6736438-14236-92540342954756/.source.volume", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Saturday 14 December 2024 11:33:31 -0500 (0:00:00.713) 0:00:40.374 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Saturday 14 December 2024 11:33:31 -0500 (0:00:00.032) 0:00:40.407 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Saturday 14 December 2024 11:33:31 -0500 (0:00:00.030) 0:00:40.437 ***** ok: [managed-node3] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Saturday 14 December 2024 11:33:32 -0500 (0:00:00.730) 0:00:41.168 ***** changed: [managed-node3] => { "changed": true, "name": "quadlet-demo-mysql-volume.service", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "sysinit.target systemd-journald.socket -.mount basic.target network-online.target system.slice", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveMemoryHigh": "3698237440", "EffectiveMemoryMax": "3698237440", "EffectiveTasksMax": "22361", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3214524416", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "[not set]", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "[not set]", "MemorySwapMax": "infinity", "MemorySwapPeak": "[not set]", "MemoryZSwapCurrent": "[not set]", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "sysinit.target system.slice -.mount", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.volume", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22361", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "Wants": "network-online.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Saturday 14 December 2024 11:33:32 -0500 (0:00:00.654) 0:00:41.822 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_service_started is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 14 December 2024 11:33:32 -0500 (0:00:00.055) 0:00:41.878 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 14 December 2024 11:33:32 -0500 (0:00:00.139) 0:00:42.018 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 14 December 2024 11:33:33 -0500 (0:00:00.065) 0:00:42.084 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 14 December 2024 11:33:33 -0500 (0:00:00.055) 0:00:42.139 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 14 December 2024 11:33:33 -0500 (0:00:00.083) 0:00:42.223 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 14 December 2024 11:33:33 -0500 (0:00:00.094) 0:00:42.317 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 14 December 2024 11:33:33 -0500 (0:00:00.060) 0:00:42.378 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 14 December 2024 11:33:33 -0500 (0:00:00.055) 0:00:42.434 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 14 December 2024 11:33:33 -0500 (0:00:00.069) 0:00:42.503 ***** ok: [managed-node3] => { "changed": false, "stat": { "atime": 1734193882.6705027, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1734193862.9755826, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 8859182, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "2878164177", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 14 December 2024 11:33:33 -0500 (0:00:00.468) 0:00:42.972 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:44 Saturday 14 December 2024 11:33:33 -0500 (0:00:00.033) 0:00:43.005 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Saturday 14 December 2024 11:33:33 -0500 (0:00:00.033) 0:00:43.038 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:62 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.028) 0:00:43.067 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:67 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.030) 0:00:43.097 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:72 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.029) 0:00:43.127 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:82 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.030) 0:00:43.157 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:89 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.029) 0:00:43.187 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.030) 0:00:43.217 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.052) 0:00:43.269 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.039) 0:00:43.309 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.040) 0:00:43.350 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:105 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.124) 0:00:43.474 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:112 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.061) 0:00:43.536 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:116 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.051) 0:00:43.587 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.109) 0:00:43.697 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.086) 0:00:43.784 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.045) 0:00:43.830 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.050) 0:00:43.881 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Saturday 14 December 2024 11:33:34 -0500 (0:00:00.043) 0:00:43.925 ***** changed: [managed-node3] => (item=/tmp/quadlet_demo) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/quadlet_demo", "mode": "0777", "owner": "root", "path": "/tmp/quadlet_demo", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Saturday 14 December 2024 11:33:35 -0500 (0:00:00.464) 0:00:44.389 ***** changed: [managed-node3] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Saturday 14 December 2024 11:33:42 -0500 (0:00:07.252) 0:00:51.642 ***** ok: [managed-node3] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 67, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Saturday 14 December 2024 11:33:42 -0500 (0:00:00.385) 0:00:52.028 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_quadlet_file_src | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Saturday 14 December 2024 11:33:43 -0500 (0:00:00.035) 0:00:52.063 ***** changed: [managed-node3] => { "changed": true, "checksum": "ca62b2ad3cc9afb5b5371ebbf797b9bc4fd7edd4", "dest": "/etc/containers/systemd/quadlet-demo-mysql.container", "gid": 0, "group": "root", "md5sum": "341b473056d2a5dfa35970b0d2e23a5d", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 363, "src": "/root/.ansible/tmp/ansible-tmp-1734194023.0743153-14590-125136684732491/.source.container", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Saturday 14 December 2024 11:33:43 -0500 (0:00:00.728) 0:00:52.792 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_copy_content is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Saturday 14 December 2024 11:33:43 -0500 (0:00:00.036) 0:00:52.829 ***** ok: [managed-node3] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Saturday 14 December 2024 11:33:44 -0500 (0:00:00.729) 0:00:53.558 ***** changed: [managed-node3] => { "changed": true, "name": "quadlet-demo-mysql.service", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "systemd-journald.socket -.mount network-online.target quadlet-demo-network.service sysinit.target system.slice tmp.mount quadlet-demo-mysql-volume.service basic.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "yes", "DelegateControllers": "cpu cpuset io memory pids", "Description": "quadlet-demo-mysql.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveMemoryHigh": "3698237440", "EffectiveMemoryMax": "3698237440", "EffectiveTasksMax": "22361", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; flags=ignore-failure ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo-mysql.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "3041705984", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "[not set]", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "[not set]", "MemorySwapMax": "infinity", "MemorySwapPeak": "[not set]", "MemoryZSwapCurrent": "[not set]", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMPolicy": "continue", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "quadlet-demo-mysql-volume.service -.mount system.slice sysinit.target quadlet-demo-network.service", "RequiresMountsFor": "/run/containers /tmp/quadlet_demo", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22361", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-online.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Saturday 14 December 2024 11:33:45 -0500 (0:00:00.951) 0:00:54.509 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_service_started is changed", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 14 December 2024 11:33:45 -0500 (0:00:00.036) 0:00:54.546 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_file_src": "envoy-proxy-configmap.yml", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 14 December 2024 11:33:45 -0500 (0:00:00.052) 0:00:54.598 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 14 December 2024 11:33:45 -0500 (0:00:00.038) 0:00:54.637 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 14 December 2024 11:33:45 -0500 (0:00:00.032) 0:00:54.669 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_name": "envoy-proxy-configmap", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 14 December 2024 11:33:45 -0500 (0:00:00.045) 0:00:54.714 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 14 December 2024 11:33:45 -0500 (0:00:00.056) 0:00:54.771 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 14 December 2024 11:33:45 -0500 (0:00:00.042) 0:00:54.813 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 14 December 2024 11:33:45 -0500 (0:00:00.052) 0:00:54.866 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 14 December 2024 11:33:45 -0500 (0:00:00.066) 0:00:54.932 ***** ok: [managed-node3] => { "changed": false, "stat": { "atime": 1734193882.6705027, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1734193862.9755826, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 8859182, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "2878164177", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.466) 0:00:55.399 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:44 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.033) 0:00:55.432 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.045) 0:00:55.478 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:62 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.104) 0:00:55.582 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:67 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.032) 0:00:55.614 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:72 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.035) 0:00:55.650 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:82 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.035) 0:00:55.685 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:89 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.034) 0:00:55.720 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.028) 0:00:55.748 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.052) 0:00:55.801 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.032) 0:00:55.833 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.032) 0:00:55.866 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:105 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.081) 0:00:55.947 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:112 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.048) 0:00:55.996 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:116 Saturday 14 December 2024 11:33:46 -0500 (0:00:00.034) 0:00:56.031 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Saturday 14 December 2024 11:33:47 -0500 (0:00:00.072) 0:00:56.103 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:33:47 -0500 (0:00:00.049) 0:00:56.152 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:33:47 -0500 (0:00:00.030) 0:00:56.182 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:33:47 -0500 (0:00:00.031) 0:00:56.214 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Saturday 14 December 2024 11:33:47 -0500 (0:00:00.038) 0:00:56.253 ***** skipping: [managed-node3] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Saturday 14 December 2024 11:33:47 -0500 (0:00:00.040) 0:00:56.293 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Saturday 14 December 2024 11:33:47 -0500 (0:00:00.047) 0:00:56.341 ***** ok: [managed-node3] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 103, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Saturday 14 December 2024 11:33:47 -0500 (0:00:00.460) 0:00:56.801 ***** changed: [managed-node3] => { "changed": true, "checksum": "d681c7d56f912150d041873e880818b22a90c188", "dest": "/etc/containers/systemd/envoy-proxy-configmap.yml", "gid": 0, "group": "root", "md5sum": "aec75d972c231aac004e1338934544cf", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 2102, "src": "/root/.ansible/tmp/ansible-tmp-1734194027.8082154-14774-224715775555281/.source.yml", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Saturday 14 December 2024 11:33:48 -0500 (0:00:00.810) 0:00:57.612 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Saturday 14 December 2024 11:33:48 -0500 (0:00:00.048) 0:00:57.660 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Saturday 14 December 2024 11:33:48 -0500 (0:00:00.052) 0:00:57.712 ***** ok: [managed-node3] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Saturday 14 December 2024 11:33:49 -0500 (0:00:00.831) 0:00:58.543 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Saturday 14 December 2024 11:33:49 -0500 (0:00:00.041) 0:00:58.585 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 14 December 2024 11:33:49 -0500 (0:00:00.037) 0:00:58.622 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n", "__podman_quadlet_template_src": "quadlet-demo.yml.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 14 December 2024 11:33:49 -0500 (0:00:00.079) 0:00:58.702 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 14 December 2024 11:33:49 -0500 (0:00:00.039) 0:00:58.742 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 14 December 2024 11:33:49 -0500 (0:00:00.033) 0:00:58.775 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 14 December 2024 11:33:49 -0500 (0:00:00.056) 0:00:58.831 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 14 December 2024 11:33:49 -0500 (0:00:00.080) 0:00:58.912 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 14 December 2024 11:33:49 -0500 (0:00:00.041) 0:00:58.953 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 14 December 2024 11:33:49 -0500 (0:00:00.041) 0:00:58.994 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 14 December 2024 11:33:50 -0500 (0:00:00.047) 0:00:59.042 ***** ok: [managed-node3] => { "changed": false, "stat": { "atime": 1734193882.6705027, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1734193862.9755826, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 8859182, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "2878164177", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 14 December 2024 11:33:50 -0500 (0:00:00.399) 0:00:59.442 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:44 Saturday 14 December 2024 11:33:50 -0500 (0:00:00.048) 0:00:59.490 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Saturday 14 December 2024 11:33:50 -0500 (0:00:00.046) 0:00:59.537 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:62 Saturday 14 December 2024 11:33:50 -0500 (0:00:00.048) 0:00:59.585 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:67 Saturday 14 December 2024 11:33:50 -0500 (0:00:00.046) 0:00:59.632 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:72 Saturday 14 December 2024 11:33:50 -0500 (0:00:00.047) 0:00:59.680 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:82 Saturday 14 December 2024 11:33:50 -0500 (0:00:00.047) 0:00:59.727 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:89 Saturday 14 December 2024 11:33:50 -0500 (0:00:00.104) 0:00:59.831 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 14 December 2024 11:33:50 -0500 (0:00:00.048) 0:00:59.880 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 14 December 2024 11:33:50 -0500 (0:00:00.088) 0:00:59.969 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 14 December 2024 11:33:50 -0500 (0:00:00.052) 0:01:00.021 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 14 December 2024 11:33:51 -0500 (0:00:00.052) 0:01:00.073 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:105 Saturday 14 December 2024 11:33:51 -0500 (0:00:00.141) 0:01:00.215 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:112 Saturday 14 December 2024 11:33:51 -0500 (0:00:00.043) 0:01:00.258 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:116 Saturday 14 December 2024 11:33:51 -0500 (0:00:00.035) 0:01:00.294 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Saturday 14 December 2024 11:33:51 -0500 (0:00:00.076) 0:01:00.371 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:33:51 -0500 (0:00:00.049) 0:01:00.420 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:33:51 -0500 (0:00:00.029) 0:01:00.450 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:33:51 -0500 (0:00:00.031) 0:01:00.482 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Saturday 14 December 2024 11:33:51 -0500 (0:00:00.029) 0:01:00.511 ***** skipping: [managed-node3] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Saturday 14 December 2024 11:33:51 -0500 (0:00:00.032) 0:01:00.544 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Saturday 14 December 2024 11:33:51 -0500 (0:00:00.042) 0:01:00.587 ***** ok: [managed-node3] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 136, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Saturday 14 December 2024 11:33:52 -0500 (0:00:00.591) 0:01:01.178 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_quadlet_file_src | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Saturday 14 December 2024 11:33:52 -0500 (0:00:00.051) 0:01:01.229 ***** changed: [managed-node3] => { "changed": true, "checksum": "998dccde0483b1654327a46ddd89cbaa47650370", "dest": "/etc/containers/systemd/quadlet-demo.yml", "gid": 0, "group": "root", "md5sum": "fd890594adfc24339cb9cdc5e7b19a66", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 1605, "src": "/root/.ansible/tmp/ansible-tmp-1734194032.2620435-14941-256399581694956/.source.yml", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Saturday 14 December 2024 11:33:53 -0500 (0:00:01.044) 0:01:02.274 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_copy_content is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Saturday 14 December 2024 11:33:53 -0500 (0:00:00.029) 0:01:02.303 ***** ok: [managed-node3] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Saturday 14 December 2024 11:33:54 -0500 (0:00:00.793) 0:01:03.097 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Saturday 14 December 2024 11:33:54 -0500 (0:00:00.034) 0:01:03.132 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 14 December 2024 11:33:54 -0500 (0:00:00.037) 0:01:03.170 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 14 December 2024 11:33:54 -0500 (0:00:00.043) 0:01:03.213 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 14 December 2024 11:33:54 -0500 (0:00:00.053) 0:01:03.267 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 14 December 2024 11:33:54 -0500 (0:00:00.046) 0:01:03.314 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 14 December 2024 11:33:54 -0500 (0:00:00.079) 0:01:03.393 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 14 December 2024 11:33:54 -0500 (0:00:00.098) 0:01:03.492 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 14 December 2024 11:33:54 -0500 (0:00:00.052) 0:01:03.544 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 14 December 2024 11:33:54 -0500 (0:00:00.057) 0:01:03.602 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 14 December 2024 11:33:54 -0500 (0:00:00.068) 0:01:03.670 ***** ok: [managed-node3] => { "changed": false, "stat": { "atime": 1734193882.6705027, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1734193862.9755826, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 8859182, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "2878164177", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.400) 0:01:04.071 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:44 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.028) 0:01:04.100 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.030) 0:01:04.131 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:62 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.030) 0:01:04.161 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:67 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.033) 0:01:04.194 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:72 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.045) 0:01:04.240 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:82 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.051) 0:01:04.291 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:89 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.033) 0:01:04.325 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.035) 0:01:04.361 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": [ "quadlet-demo.yml" ], "__podman_service_name": "quadlet-demo.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.062) 0:01:04.423 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.074) 0:01:04.498 ***** ok: [managed-node3] => { "changed": false, "content": "LS0tCmFwaVZlcnNpb246IHYxCmtpbmQ6IFBlcnNpc3RlbnRWb2x1bWVDbGFpbQptZXRhZGF0YToKICBuYW1lOiB3cC1wdi1jbGFpbQogIGxhYmVsczoKICAgIGFwcDogd29yZHByZXNzCnNwZWM6CiAgYWNjZXNzTW9kZXM6CiAgLSBSZWFkV3JpdGVPbmNlCiAgcmVzb3VyY2VzOgogICAgcmVxdWVzdHM6CiAgICAgIHN0b3JhZ2U6IDIwR2kKLS0tCmFwaVZlcnNpb246IHYxCmtpbmQ6IFBvZAptZXRhZGF0YToKICBuYW1lOiBxdWFkbGV0LWRlbW8Kc3BlYzoKICBjb250YWluZXJzOgogIC0gbmFtZTogd29yZHByZXNzCiAgICBpbWFnZTogcXVheS5pby9saW51eC1zeXN0ZW0tcm9sZXMvd29yZHByZXNzOjQuOC1hcGFjaGUKICAgIGVudjoKICAgIC0gbmFtZTogV09SRFBSRVNTX0RCX0hPU1QKICAgICAgdmFsdWU6IHF1YWRsZXQtZGVtby1teXNxbAogICAgLSBuYW1lOiBXT1JEUFJFU1NfREJfUEFTU1dPUkQKICAgICAgdmFsdWVGcm9tOgogICAgICAgIHNlY3JldEtleVJlZjoKICAgICAgICAgIG5hbWU6IG15c3FsLXJvb3QtcGFzc3dvcmQta3ViZQogICAgICAgICAga2V5OiBwYXNzd29yZAogICAgdm9sdW1lTW91bnRzOgogICAgLSBuYW1lOiB3b3JkcHJlc3MtcGVyc2lzdGVudC1zdG9yYWdlCiAgICAgIG1vdW50UGF0aDogL3Zhci93d3cvaHRtbAogICAgcmVzb3VyY2VzOgogICAgICByZXF1ZXN0czoKICAgICAgICBtZW1vcnk6ICI2NE1pIgogICAgICAgIGNwdTogIjI1MG0iCiAgICAgIGxpbWl0czoKICAgICAgICBtZW1vcnk6ICIxMjhNaSIKICAgICAgICBjcHU6ICI1MDBtIgogIC0gbmFtZTogZW52b3kKICAgIGltYWdlOiBxdWF5LmlvL2xpbnV4LXN5c3RlbS1yb2xlcy9lbnZveXByb3h5OnYxLjI1LjAKICAgIHZvbHVtZU1vdW50czoKICAgIC0gbmFtZTogY29uZmlnLXZvbHVtZQogICAgICBtb3VudFBhdGg6IC9ldGMvZW52b3kKICAgIC0gbmFtZTogY2VydGlmaWNhdGVzCiAgICAgIG1vdW50UGF0aDogL2V0Yy9lbnZveS1jZXJ0aWZpY2F0ZXMKICAgIGVudjoKICAgIC0gbmFtZTogRU5WT1lfVUlECiAgICAgIHZhbHVlOiAiMCIKICAgIHJlc291cmNlczoKICAgICAgcmVxdWVzdHM6CiAgICAgICAgbWVtb3J5OiAiNjRNaSIKICAgICAgICBjcHU6ICIyNTBtIgogICAgICBsaW1pdHM6CiAgICAgICAgbWVtb3J5OiAiMTI4TWkiCiAgICAgICAgY3B1OiAiNTAwbSIKICB2b2x1bWVzOgogIC0gbmFtZTogY29uZmlnLXZvbHVtZQogICAgY29uZmlnTWFwOgogICAgICBuYW1lOiBlbnZveS1wcm94eS1jb25maWcKICAtIG5hbWU6IGNlcnRpZmljYXRlcwogICAgc2VjcmV0OgogICAgICBzZWNyZXROYW1lOiBlbnZveS1jZXJ0aWZpY2F0ZXMKICAtIG5hbWU6IHdvcmRwcmVzcy1wZXJzaXN0ZW50LXN0b3JhZ2UKICAgIHBlcnNpc3RlbnRWb2x1bWVDbGFpbToKICAgICAgY2xhaW1OYW1lOiB3cC1wdi1jbGFpbQogIC0gbmFtZTogd3d3ICAjIG5vdCB1c2VkIC0gZm9yIHRlc3RpbmcgaG9zdHBhdGgKICAgIGhvc3RQYXRoOgogICAgICBwYXRoOiAvdG1wL2h0dHBkMwogIC0gbmFtZTogY3JlYXRlICAjIG5vdCB1c2VkIC0gZm9yIHRlc3RpbmcgaG9zdHBhdGgKICAgIGhvc3RQYXRoOgogICAgICBwYXRoOiAvdG1wL2h0dHBkMy1jcmVhdGUK", "encoding": "base64", "source": "/etc/containers/systemd/quadlet-demo.yml" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.376) 0:01:04.874 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/wordpress:4.8-apache", "quay.io/linux-system-roles/envoyproxy:v1.25.0" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube", "__podman_volumes": [ "/tmp/httpd3", "/tmp/httpd3-create" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:105 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.088) 0:01:04.963 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:112 Saturday 14 December 2024 11:33:55 -0500 (0:00:00.041) 0:01:05.004 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:116 Saturday 14 December 2024 11:33:56 -0500 (0:00:00.046) 0:01:05.051 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Saturday 14 December 2024 11:33:56 -0500 (0:00:00.107) 0:01:05.159 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:33:56 -0500 (0:00:00.081) 0:01:05.240 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:33:56 -0500 (0:00:00.052) 0:01:05.292 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:33:56 -0500 (0:00:00.055) 0:01:05.348 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Saturday 14 December 2024 11:33:56 -0500 (0:00:00.050) 0:01:05.398 ***** changed: [managed-node3] => (item=/tmp/httpd3) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/httpd3", "mode": "0755", "owner": "root", "path": "/tmp/httpd3", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } changed: [managed-node3] => (item=/tmp/httpd3-create) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/httpd3-create", "mode": "0755", "owner": "root", "path": "/tmp/httpd3-create", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Saturday 14 December 2024 11:33:57 -0500 (0:00:00.797) 0:01:06.196 ***** changed: [managed-node3] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [managed-node3] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Saturday 14 December 2024 11:34:15 -0500 (0:00:18.123) 0:01:24.320 ***** ok: [managed-node3] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 160, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Saturday 14 December 2024 11:34:15 -0500 (0:00:00.394) 0:01:24.714 ***** changed: [managed-node3] => { "changed": true, "checksum": "7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7", "dest": "/etc/containers/systemd/quadlet-demo.kube", "gid": 0, "group": "root", "md5sum": "da53c88f92b68b0487aa209f795b6bb3", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 456, "src": "/root/.ansible/tmp/ansible-tmp-1734194055.7196329-15462-153591703695286/.source.kube", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Saturday 14 December 2024 11:34:16 -0500 (0:00:00.731) 0:01:25.445 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Saturday 14 December 2024 11:34:16 -0500 (0:00:00.032) 0:01:25.478 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_copy_file is skipped", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Saturday 14 December 2024 11:34:16 -0500 (0:00:00.029) 0:01:25.507 ***** ok: [managed-node3] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Saturday 14 December 2024 11:34:17 -0500 (0:00:00.749) 0:01:26.256 ***** fatal: [managed-node3]: FAILED! => { "changed": false } MSG: Unable to start service quadlet-demo.service: Job for quadlet-demo.service failed because the control process exited with error code. See "systemctl status quadlet-demo.service" and "journalctl -xeu quadlet-demo.service" for details. TASK [Dump journal] ************************************************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:142 Saturday 14 December 2024 11:34:17 -0500 (0:00:00.636) 0:01:26.893 ***** fatal: [managed-node3]: FAILED! => { "changed": false, "cmd": [ "journalctl", "-ex" ], "delta": "0:00:00.028028", "end": "2024-12-14 11:34:18.183902", "failed_when_result": true, "rc": 0, "start": "2024-12-14 11:34:18.155874" } STDOUT: Dec 14 11:26:32 localhost systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount has successfully entered the 'dead' state. Dec 14 11:26:32 localhost systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-credentials-systemd\x2dsysctl.service.mount has successfully entered the 'dead' state. Dec 14 11:26:32 localhost systemd[1]: run-credentials-systemd\x2dudev\x2dload\x2dcredentials.service.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-credentials-systemd\x2dudev\x2dload\x2dcredentials.service.mount has successfully entered the 'dead' state. Dec 14 11:26:32 localhost systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... ░░ Subject: A start job for unit sys-fs-fuse-connections.mount has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sys-fs-fuse-connections.mount has begun execution. ░░ ░░ The job identifier is 191. Dec 14 11:26:32 localhost systemd[1]: Mounting var-lib-nfs-rpc_pipefs.mount - RPC Pipe File System... ░░ Subject: A start job for unit var-lib-nfs-rpc_pipefs.mount has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit var-lib-nfs-rpc_pipefs.mount has begun execution. ░░ ░░ The job identifier is 242. Dec 14 11:26:32 localhost systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. ░░ Subject: A start job for unit sys-fs-fuse-connections.mount has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sys-fs-fuse-connections.mount has finished successfully. ░░ ░░ The job identifier is 191. Dec 14 11:26:32 localhost systemd[1]: etc-machine\x2did.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit etc-machine\x2did.mount has successfully entered the 'dead' state. Dec 14 11:26:32 localhost systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. ░░ Subject: A start job for unit systemd-machine-id-commit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-machine-id-commit.service has finished successfully. ░░ ░░ The job identifier is 148. Dec 14 11:26:32 localhost systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. ░░ Subject: A start job for unit systemd-udevd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-udevd.service has finished successfully. ░░ ░░ The job identifier is 151. Dec 14 11:26:32 localhost systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... ░░ Subject: A start job for unit modprobe@configfs.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@configfs.service has begun execution. ░░ ░░ The job identifier is 294. Dec 14 11:26:32 localhost systemd[1]: modprobe@configfs.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit modprobe@configfs.service has successfully entered the 'dead' state. Dec 14 11:26:32 localhost systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. ░░ Subject: A start job for unit modprobe@configfs.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit modprobe@configfs.service has finished successfully. ░░ ░░ The job identifier is 294. Dec 14 11:26:32 localhost systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. ░░ Subject: A start job for unit dev-ttyS0.device has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dev-ttyS0.device has finished successfully. ░░ ░░ The job identifier is 229. Dec 14 11:26:32 localhost kernel: RPC: Registered named UNIX socket transport module. Dec 14 11:26:32 localhost kernel: RPC: Registered udp transport module. Dec 14 11:26:32 localhost kernel: RPC: Registered tcp transport module. Dec 14 11:26:32 localhost kernel: RPC: Registered tcp-with-tls transport module. Dec 14 11:26:32 localhost kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Dec 14 11:26:32 localhost systemd[1]: Mounted var-lib-nfs-rpc_pipefs.mount - RPC Pipe File System. ░░ Subject: A start job for unit var-lib-nfs-rpc_pipefs.mount has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit var-lib-nfs-rpc_pipefs.mount has finished successfully. ░░ ░░ The job identifier is 242. Dec 14 11:26:32 localhost systemd[1]: Reached target rpc_pipefs.target. ░░ Subject: A start job for unit rpc_pipefs.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc_pipefs.target has finished successfully. ░░ ░░ The job identifier is 241. Dec 14 11:26:32 localhost (udev-worker)[581]: Network interface NamePolicy= disabled on kernel command line. Dec 14 11:26:32 localhost kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 655360 ms ovfl timer Dec 14 11:26:32 localhost systemd[1]: Started rpcbind.service - RPC Bind. ░░ Subject: A start job for unit rpcbind.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpcbind.service has finished successfully. ░░ ░░ The job identifier is 251. Dec 14 11:26:32 localhost kernel: input: PC Speaker as /devices/platform/pcspkr/input/input5 Dec 14 11:26:32 localhost augenrules[563]: /sbin/augenrules: No change Dec 14 11:26:32 localhost augenrules[617]: No rules Dec 14 11:26:32 localhost augenrules[617]: enabled 0 Dec 14 11:26:32 localhost augenrules[617]: failure 1 Dec 14 11:26:32 localhost augenrules[617]: pid 0 Dec 14 11:26:32 localhost augenrules[617]: rate_limit 0 Dec 14 11:26:32 localhost augenrules[617]: backlog_limit 8192 Dec 14 11:26:32 localhost augenrules[617]: lost 0 Dec 14 11:26:32 localhost augenrules[617]: backlog 0 Dec 14 11:26:32 localhost augenrules[617]: backlog_wait_time 60000 Dec 14 11:26:32 localhost augenrules[617]: backlog_wait_time_actual 0 Dec 14 11:26:32 localhost augenrules[617]: enabled 0 Dec 14 11:26:32 localhost augenrules[617]: failure 1 Dec 14 11:26:32 localhost augenrules[617]: pid 0 Dec 14 11:26:32 localhost augenrules[617]: rate_limit 0 Dec 14 11:26:32 localhost augenrules[617]: backlog_limit 8192 Dec 14 11:26:32 localhost augenrules[617]: lost 0 Dec 14 11:26:32 localhost augenrules[617]: backlog 0 Dec 14 11:26:32 localhost augenrules[617]: backlog_wait_time 60000 Dec 14 11:26:32 localhost augenrules[617]: backlog_wait_time_actual 0 Dec 14 11:26:32 localhost augenrules[617]: enabled 0 Dec 14 11:26:32 localhost augenrules[617]: failure 1 Dec 14 11:26:32 localhost augenrules[617]: pid 0 Dec 14 11:26:32 localhost augenrules[617]: rate_limit 0 Dec 14 11:26:32 localhost augenrules[617]: backlog_limit 8192 Dec 14 11:26:32 localhost augenrules[617]: lost 0 Dec 14 11:26:32 localhost augenrules[617]: backlog 0 Dec 14 11:26:32 localhost augenrules[617]: backlog_wait_time 60000 Dec 14 11:26:32 localhost augenrules[617]: backlog_wait_time_actual 0 Dec 14 11:26:32 localhost systemd[1]: audit-rules.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit audit-rules.service has successfully entered the 'dead' state. Dec 14 11:26:32 localhost systemd[1]: Finished audit-rules.service - Load Audit Rules. ░░ Subject: A start job for unit audit-rules.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit audit-rules.service has finished successfully. ░░ ░░ The job identifier is 248. Dec 14 11:26:32 localhost systemd[1]: Starting auditd.service - Security Audit Logging Service... ░░ Subject: A start job for unit auditd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has begun execution. ░░ ░░ The job identifier is 247. Dec 14 11:26:33 localhost auditd[628]: No plugins found, not dispatching events Dec 14 11:26:33 localhost auditd[628]: Init complete, auditd 4.0 listening for events (startup state enable) Dec 14 11:26:33 localhost systemd[1]: Started auditd.service - Security Audit Logging Service. ░░ Subject: A start job for unit auditd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has finished successfully. ░░ ░░ The job identifier is 247. Dec 14 11:26:33 localhost kernel: cirrus 0000:00:02.0: vgaarb: deactivate vga console Dec 14 11:26:33 localhost kernel: Console: switching to colour dummy device 80x25 Dec 14 11:26:33 localhost kernel: [drm] Initialized cirrus 2.0.0 for 0000:00:02.0 on minor 0 Dec 14 11:26:33 localhost kernel: fbcon: cirrusdrmfb (fb0) is primary device Dec 14 11:26:33 localhost kernel: Console: switching to colour frame buffer device 128x48 Dec 14 11:26:33 localhost kernel: cirrus 0000:00:02.0: [drm] fb0: cirrusdrmfb frame buffer device Dec 14 11:26:33 localhost systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... ░░ Subject: A start job for unit systemd-update-utmp.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp.service has begun execution. ░░ ░░ The job identifier is 236. Dec 14 11:26:33 localhost systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... ░░ Subject: A start job for unit systemd-vconsole-setup.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-vconsole-setup.service has begun execution. ░░ ░░ The job identifier is 300. Dec 14 11:26:33 localhost systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. ░░ Subject: A start job for unit systemd-update-utmp.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp.service has finished successfully. ░░ ░░ The job identifier is 236. Dec 14 11:26:33 localhost kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Dec 14 11:26:33 localhost systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-vconsole-setup.service has successfully entered the 'dead' state. Dec 14 11:26:33 localhost systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. ░░ Subject: A stop job for unit systemd-vconsole-setup.service has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit systemd-vconsole-setup.service has finished. ░░ ░░ The job identifier is 300 and the job result is done. Dec 14 11:26:33 localhost systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... ░░ Subject: A start job for unit systemd-vconsole-setup.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-vconsole-setup.service has begun execution. ░░ ░░ The job identifier is 300. Dec 14 11:26:33 localhost systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. ░░ Subject: A start job for unit systemd-vconsole-setup.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-vconsole-setup.service has finished successfully. ░░ ░░ The job identifier is 300. Dec 14 11:26:33 localhost systemd[1]: Reached target sysinit.target - System Initialization. ░░ Subject: A start job for unit sysinit.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sysinit.target has finished successfully. ░░ ░░ The job identifier is 123. Dec 14 11:26:33 localhost systemd[1]: Started dnf-makecache.timer - dnf makecache --timer. ░░ Subject: A start job for unit dnf-makecache.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dnf-makecache.timer has finished successfully. ░░ ░░ The job identifier is 204. Dec 14 11:26:33 localhost systemd[1]: Started fstrim.timer - Discard unused filesystem blocks once a week. ░░ Subject: A start job for unit fstrim.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit fstrim.timer has finished successfully. ░░ ░░ The job identifier is 202. Dec 14 11:26:33 localhost systemd[1]: Started logrotate.timer - Daily rotation of log files. ░░ Subject: A start job for unit logrotate.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.timer has finished successfully. ░░ ░░ The job identifier is 203. Dec 14 11:26:33 localhost systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. ░░ Subject: A start job for unit systemd-tmpfiles-clean.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-clean.timer has finished successfully. ░░ ░░ The job identifier is 211. Dec 14 11:26:33 localhost systemd[1]: Reached target timers.target - Timer Units. ░░ Subject: A start job for unit timers.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit timers.target has finished successfully. ░░ ░░ The job identifier is 201. Dec 14 11:26:33 localhost systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. ░░ Subject: A start job for unit dbus.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus.socket has finished successfully. ░░ ░░ The job identifier is 208. Dec 14 11:26:33 localhost systemd[1]: Listening on pcscd.socket - PC/SC Smart Card Daemon Activation Socket. ░░ Subject: A start job for unit pcscd.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit pcscd.socket has finished successfully. ░░ ░░ The job identifier is 214. Dec 14 11:26:33 localhost systemd[1]: Listening on sssd-kcm.socket - SSSD Kerberos Cache Manager responder socket. ░░ Subject: A start job for unit sssd-kcm.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sssd-kcm.socket has finished successfully. ░░ ░░ The job identifier is 218. Dec 14 11:26:33 localhost systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. ░░ Subject: A start job for unit systemd-hostnamed.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.socket has finished successfully. ░░ ░░ The job identifier is 221. Dec 14 11:26:33 localhost systemd[1]: Reached target sockets.target - Socket Units. ░░ Subject: A start job for unit sockets.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sockets.target has finished successfully. ░░ ░░ The job identifier is 213. Dec 14 11:26:33 localhost systemd[1]: Starting dbus-broker.service - D-Bus System Message Bus... ░░ Subject: A start job for unit dbus-broker.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus-broker.service has begun execution. ░░ ░░ The job identifier is 209. Dec 14 11:26:33 localhost systemd[1]: systemd-pcrphase-sysinit.service - TPM PCR Barrier (Initialization) was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ░░ Subject: A start job for unit systemd-pcrphase-sysinit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-pcrphase-sysinit.service has finished successfully. ░░ ░░ The job identifier is 198. Dec 14 11:26:33 localhost systemd[1]: Started dbus-broker.service - D-Bus System Message Bus. ░░ Subject: A start job for unit dbus-broker.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus-broker.service has finished successfully. ░░ ░░ The job identifier is 209. Dec 14 11:26:33 localhost systemd[1]: Reached target basic.target - Basic System. ░░ Subject: A start job for unit basic.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit basic.target has finished successfully. ░░ ░░ The job identifier is 122. Dec 14 11:26:33 localhost dbus-broker-launch[646]: Ready Dec 14 11:26:33 localhost systemd[1]: Starting chronyd.service - NTP client/server... ░░ Subject: A start job for unit chronyd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has begun execution. ░░ ░░ The job identifier is 256. Dec 14 11:26:33 localhost systemd[1]: Starting cloud-init-local.service - Initial cloud-init job (pre-networking)... ░░ Subject: A start job for unit cloud-init-local.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init-local.service has begun execution. ░░ ░░ The job identifier is 268. Dec 14 11:26:33 localhost systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... ░░ Subject: A start job for unit dracut-shutdown.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dracut-shutdown.service has begun execution. ░░ ░░ The job identifier is 173. Dec 14 11:26:33 localhost systemd[1]: Started irqbalance.service - irqbalance daemon. ░░ Subject: A start job for unit irqbalance.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit irqbalance.service has finished successfully. ░░ ░░ The job identifier is 263. Dec 14 11:26:33 localhost systemd[1]: Started rngd.service - Hardware RNG Entropy Gatherer Daemon. ░░ Subject: A start job for unit rngd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rngd.service has finished successfully. ░░ ░░ The job identifier is 261. Dec 14 11:26:33 localhost (qbalance)[651]: irqbalance.service: Referenced but unset environment variable evaluates to an empty string: IRQBALANCE_ARGS Dec 14 11:26:33 localhost systemd[1]: Starting rsyslog.service - System Logging Service... ░░ Subject: A start job for unit rsyslog.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rsyslog.service has begun execution. ░░ ░░ The job identifier is 259. Dec 14 11:26:33 localhost systemd[1]: ssh-host-keys-migration.service - Update OpenSSH host key permissions was skipped because of an unmet condition check (ConditionPathExists=!/var/lib/.ssh-host-keys-migration). ░░ Subject: A start job for unit ssh-host-keys-migration.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit ssh-host-keys-migration.service has finished successfully. ░░ ░░ The job identifier is 275. Dec 14 11:26:33 localhost systemd[1]: sshd-keygen@ecdsa.service - OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ecdsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ecdsa.service has finished successfully. ░░ ░░ The job identifier is 274. Dec 14 11:26:33 localhost systemd[1]: sshd-keygen@ed25519.service - OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ed25519.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ed25519.service has finished successfully. ░░ ░░ The job identifier is 273. Dec 14 11:26:33 localhost systemd[1]: sshd-keygen@rsa.service - OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@rsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@rsa.service has finished successfully. ░░ ░░ The job identifier is 271. Dec 14 11:26:33 localhost systemd[1]: Reached target sshd-keygen.target. ░░ Subject: A start job for unit sshd-keygen.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen.target has finished successfully. ░░ ░░ The job identifier is 270. Dec 14 11:26:33 localhost systemd[1]: sssd.service - System Security Services Daemon was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit sssd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sssd.service has finished successfully. ░░ ░░ The job identifier is 249. Dec 14 11:26:33 localhost systemd[1]: Reached target nss-user-lookup.target - User and Group Name Lookups. ░░ Subject: A start job for unit nss-user-lookup.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit nss-user-lookup.target has finished successfully. ░░ ░░ The job identifier is 250. Dec 14 11:26:33 localhost systemd[1]: Starting systemd-logind.service - User Login Management... ░░ Subject: A start job for unit systemd-logind.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-logind.service has begun execution. ░░ ░░ The job identifier is 253. Dec 14 11:26:33 localhost systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-credentials-systemd\x2dvconsole\x2dsetup.service.mount has successfully entered the 'dead' state. Dec 14 11:26:33 localhost systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. ░░ Subject: A start job for unit dracut-shutdown.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dracut-shutdown.service has finished successfully. ░░ ░░ The job identifier is 173. Dec 14 11:26:33 localhost systemd-logind[655]: New seat seat0. ░░ Subject: A new seat seat0 is now available ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new seat seat0 has been configured and is now available. Dec 14 11:26:33 localhost systemd-logind[655]: Watching system buttons on /dev/input/event0 (Power Button) Dec 14 11:26:33 localhost systemd-logind[655]: Watching system buttons on /dev/input/event1 (Sleep Button) Dec 14 11:26:33 localhost systemd-logind[655]: Watching system buttons on /dev/input/event2 (AT Translated Set 2 keyboard) Dec 14 11:26:33 localhost systemd[1]: Started systemd-logind.service - User Login Management. ░░ Subject: A start job for unit systemd-logind.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-logind.service has finished successfully. ░░ ░░ The job identifier is 253. Dec 14 11:26:33 localhost systemd[1]: Started rsyslog.service - System Logging Service. ░░ Subject: A start job for unit rsyslog.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rsyslog.service has finished successfully. ░░ ░░ The job identifier is 259. Dec 14 11:26:33 localhost rsyslogd[654]: imjournal: filecreatemode is not set, using default 0644 [v8.2408.0-2.el10 try https://www.rsyslog.com/e/2186 ] Dec 14 11:26:33 localhost rsyslogd[654]: [origin software="rsyslogd" swVersion="8.2408.0-2.el10" x-pid="654" x-info="https://www.rsyslog.com"] start Dec 14 11:26:33 localhost rsyslogd[654]: imjournal: journal files changed, reloading... [v8.2408.0-2.el10 try https://www.rsyslog.com/e/0 ] Dec 14 11:26:33 localhost chronyd[662]: chronyd version 4.6.1 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Dec 14 11:26:33 localhost chronyd[662]: Frequency 0.000 +/- 1000000.000 ppm read from /var/lib/chrony/drift Dec 14 11:26:33 localhost chronyd[662]: Loaded seccomp filter (level 2) Dec 14 11:26:33 localhost systemd[1]: Started chronyd.service - NTP client/server. ░░ Subject: A start job for unit chronyd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has finished successfully. ░░ ░░ The job identifier is 256. Dec 14 11:26:33 localhost rngd[653]: Disabling 7: PKCS11 Entropy generator (pkcs11) Dec 14 11:26:33 localhost rngd[653]: Disabling 5: NIST Network Entropy Beacon (nist) Dec 14 11:26:33 localhost rngd[653]: Disabling 9: Qrypt quantum entropy beacon (qrypt) Dec 14 11:26:33 localhost rngd[653]: Disabling 10: Named pipe entropy input (namedpipe) Dec 14 11:26:33 localhost rngd[653]: Initializing available sources Dec 14 11:26:33 localhost rngd[653]: [hwrng ]: Initialization Failed Dec 14 11:26:33 localhost rngd[653]: [rdrand]: Enabling RDRAND rng support Dec 14 11:26:33 localhost rngd[653]: [rdrand]: Initialized Dec 14 11:26:33 localhost rngd[653]: [jitter]: JITTER timeout set to 5 sec Dec 14 11:26:33 localhost rngd[653]: [jitter]: Initializing AES buffer Dec 14 11:26:37 localhost cloud-init[669]: Cloud-init v. 24.1.4-21.el10 running 'init-local' at Sat, 14 Dec 2024 16:26:37 +0000. Up 18.36 seconds. Dec 14 11:26:37 localhost dhcpcd[671]: dhcpcd-10.0.6 starting Dec 14 11:26:37 localhost kernel: 8021q: 802.1Q VLAN Support v1.8 Dec 14 11:26:37 localhost systemd[1]: Listening on systemd-rfkill.socket - Load/Save RF Kill Switch Status /dev/rfkill Watch. ░░ Subject: A start job for unit systemd-rfkill.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-rfkill.socket has finished successfully. ░░ ░░ The job identifier is 310. Dec 14 11:26:37 localhost kernel: cfg80211: Loading compiled-in X.509 certificates for regulatory database Dec 14 11:26:37 localhost kernel: Loaded X.509 cert 'sforshee: 00b28ddf47aef9cea7' Dec 14 11:26:37 localhost kernel: Loaded X.509 cert 'wens: 61c038651aabdcf94bd0ac7ff06c7248db18c600' Dec 14 11:26:37 localhost dhcpcd[674]: DUID 00:01:00:01:2e:f0:6e:3d:0a:ff:fb:3d:09:35 Dec 14 11:26:37 localhost dhcpcd[674]: eth0: IAID fb:3d:09:35 Dec 14 11:26:37 localhost kernel: platform regulatory.0: Direct firmware load for regulatory.db failed with error -2 Dec 14 11:26:37 localhost kernel: cfg80211: failed to load regulatory.db Dec 14 11:26:38 localhost rngd[653]: [jitter]: Unable to obtain AES key, disabling JITTER source Dec 14 11:26:38 localhost rngd[653]: [jitter]: Initialization Failed Dec 14 11:26:38 localhost rngd[653]: Process privileges have been dropped to 2:2 Dec 14 11:26:38 localhost dhcpcd[674]: eth0: soliciting a DHCP lease Dec 14 11:26:38 localhost dhcpcd[674]: eth0: offered 10.31.14.110 from 10.31.12.1 Dec 14 11:26:38 localhost dhcpcd[674]: eth0: leased 10.31.14.110 for 3600 seconds Dec 14 11:26:38 localhost dhcpcd[674]: eth0: adding route to 10.31.12.0/22 Dec 14 11:26:38 localhost dhcpcd[674]: eth0: adding default route via 10.31.12.1 Dec 14 11:26:38 localhost dhcpcd[674]: control command: /usr/sbin/dhcpcd --dumplease --ipv4only eth0 Dec 14 11:26:38 localhost systemd[1]: Starting systemd-hostnamed.service - Hostname Service... ░░ Subject: A start job for unit systemd-hostnamed.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has begun execution. ░░ ░░ The job identifier is 319. Dec 14 11:26:38 localhost systemd[1]: Started systemd-hostnamed.service - Hostname Service. ░░ Subject: A start job for unit systemd-hostnamed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has finished successfully. ░░ ░░ The job identifier is 319. Dec 14 11:26:38 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-hostnamed[696]: Hostname set to (static) Dec 14 11:26:38 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished cloud-init-local.service - Initial cloud-init job (pre-networking). ░░ Subject: A start job for unit cloud-init-local.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init-local.service has finished successfully. ░░ ░░ The job identifier is 268. Dec 14 11:26:38 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target network-pre.target - Preparation for Network. ░░ Subject: A start job for unit network-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network-pre.target has finished successfully. ░░ ░░ The job identifier is 159. Dec 14 11:26:38 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting NetworkManager.service - Network Manager... ░░ Subject: A start job for unit NetworkManager.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager.service has begun execution. ░░ ░░ The job identifier is 207. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.2519] NetworkManager (version 1.51.4-1.el10) is starting... (boot:67ab4bf4-984b-4f4d-8c19-30dde3975820) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.2521] Read config: /etc/NetworkManager/NetworkManager.conf, /etc/NetworkManager/conf.d/30-cloud-init-ip6-addr-gen-mode.conf Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3090] manager[0x5614ecd3f9f0]: monitoring kernel firmware directory '/lib/firmware'. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3122] hostname: hostname: using hostnamed Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3122] hostname: static hostname changed from (none) to "ip-10-31-14-110.us-east-1.aws.redhat.com" Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3129] dns-mgr: init: dns=default,systemd-resolved rc-manager=symlink (auto) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3133] manager[0x5614ecd3f9f0]: rfkill: Wi-Fi hardware radio set enabled Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3134] manager[0x5614ecd3f9f0]: rfkill: WWAN hardware radio set enabled Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3181] manager: rfkill: Wi-Fi enabled by radio killswitch; enabled by state file Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3182] manager: rfkill: WWAN enabled by radio killswitch; enabled by state file Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3182] manager: Networking is enabled by state file Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3213] settings: Loaded settings plugin: keyfile (internal) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 397. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3538] dhcp: init: Using DHCP client 'internal' Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3540] manager: (lo): new Loopback device (/org/freedesktop/NetworkManager/Devices/1) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3549] device (lo): state change: unmanaged -> unavailable (reason 'connection-assumed', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3566] device (lo): state change: unavailable -> disconnected (reason 'connection-assumed', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3572] device (lo): Activation: starting connection 'lo' (8e242600-3605-43e7-b815-8160476017ad) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3578] manager: (eth0): new Ethernet device (/org/freedesktop/NetworkManager/Devices/2) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3581] device (eth0): state change: unmanaged -> unavailable (reason 'managed', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started NetworkManager.service - Network Manager. ░░ Subject: A start job for unit NetworkManager.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager.service has finished successfully. ░░ ░░ The job identifier is 207. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3596] bus-manager: acquired D-Bus service "org.freedesktop.NetworkManager" Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3599] device (lo): state change: disconnected -> prepare (reason 'none', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3610] device (lo): state change: prepare -> config (reason 'none', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3612] device (lo): state change: config -> ip-config (reason 'none', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3621] device (eth0): carrier: link connected Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target network.target - Network. ░░ Subject: A start job for unit network.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network.target has finished successfully. ░░ ░░ The job identifier is 210. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3641] device (lo): state change: ip-config -> ip-check (reason 'none', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3646] device (eth0): state change: unavailable -> disconnected (reason 'carrier-changed', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3650] policy: auto-activating connection 'cloud-init eth0' (1dd9a779-d327-56e1-8454-c65e2556c12c) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3654] device (eth0): Activation: starting connection 'cloud-init eth0' (1dd9a779-d327-56e1-8454-c65e2556c12c) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3655] device (eth0): state change: disconnected -> prepare (reason 'none', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3657] manager: NetworkManager state is now CONNECTING Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3658] device (eth0): state change: prepare -> config (reason 'none', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3662] device (eth0): state change: config -> ip-config (reason 'none', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3671] dhcp4 (eth0): activation: beginning transaction (timeout in 45 seconds) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3689] dhcp4 (eth0): state changed new lease, address=10.31.14.110, acd pending Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting NetworkManager-wait-online.service - Network Manager Wait Online... ░░ Subject: A start job for unit NetworkManager-wait-online.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-wait-online.service has begun execution. ░░ ░░ The job identifier is 206. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting gssproxy.service - GSSAPI Proxy Daemon... ░░ Subject: A start job for unit gssproxy.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit gssproxy.service has begun execution. ░░ ░░ The job identifier is 244. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 397. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5143] dhcp4 (eth0): state changed new lease, address=10.31.14.110 Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5181] policy: set 'cloud-init eth0' (eth0) as default for IPv4 routing and DNS Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5904] device (lo): state change: ip-check -> secondaries (reason 'none', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5909] device (eth0): state change: ip-config -> ip-check (reason 'none', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5918] device (lo): state change: secondaries -> activated (reason 'none', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5922] device (lo): Activation: successful, device activated. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5970] device (eth0): state change: ip-check -> secondaries (reason 'none', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5976] device (eth0): state change: secondaries -> activated (reason 'none', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5980] manager: NetworkManager state is now CONNECTED_SITE Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5983] device (eth0): Activation: successful, device activated. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5987] manager: NetworkManager state is now CONNECTED_GLOBAL Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5990] manager: startup complete Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished NetworkManager-wait-online.service - Network Manager Wait Online. ░░ Subject: A start job for unit NetworkManager-wait-online.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-wait-online.service has finished successfully. ░░ ░░ The job identifier is 206. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting cloud-init.service - Initial cloud-init job (metadata service crawler)... ░░ Subject: A start job for unit cloud-init.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.service has begun execution. ░░ ░░ The job identifier is 267. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started gssproxy.service - GSSAPI Proxy Daemon. ░░ Subject: A start job for unit gssproxy.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit gssproxy.service has finished successfully. ░░ ░░ The job identifier is 244. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: rpc-gssd.service - RPC security service for NFS client and server was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). ░░ Subject: A start job for unit rpc-gssd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-gssd.service has finished successfully. ░░ ░░ The job identifier is 240. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target nfs-client.target - NFS client services. ░░ Subject: A start job for unit nfs-client.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit nfs-client.target has finished successfully. ░░ ░░ The job identifier is 238. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. ░░ Subject: A start job for unit remote-fs-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-fs-pre.target has finished successfully. ░░ ░░ The job identifier is 245. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. ░░ Subject: A start job for unit remote-cryptsetup.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-cryptsetup.target has finished successfully. ░░ ░░ The job identifier is 234. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target remote-fs.target - Remote File Systems. ░░ Subject: A start job for unit remote-fs.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-fs.target has finished successfully. ░░ ░░ The job identifier is 277. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: systemd-pcrphase.service - TPM PCR Barrier (User) was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ░░ Subject: A start job for unit systemd-pcrphase.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-pcrphase.service has finished successfully. ░░ ░░ The job identifier is 196. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com chronyd[662]: Added source 10.11.160.238 Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com chronyd[662]: Added source 10.18.100.10 Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com chronyd[662]: Added source 10.2.32.37 Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com chronyd[662]: Added source 10.2.32.38 Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Cloud-init v. 24.1.4-21.el10 running 'init' at Sat, 14 Dec 2024 16:26:39 +0000. Up 21.09 seconds. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: ++++++++++++++++++++++++++++++++++++++Net device info++++++++++++++++++++++++++++++++++++++ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +--------+------+----------------------------+---------------+--------+-------------------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | Device | Up | Address | Mask | Scope | Hw-Address | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +--------+------+----------------------------+---------------+--------+-------------------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | eth0 | True | 10.31.14.110 | 255.255.252.0 | global | 0a:ff:fb:3d:09:35 | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | eth0 | True | fe80::8ff:fbff:fe3d:935/64 | . | link | 0a:ff:fb:3d:09:35 | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | lo | True | 127.0.0.1 | 255.0.0.0 | host | . | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | lo | True | ::1/128 | . | host | . | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +--------+------+----------------------------+---------------+--------+-------------------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: ++++++++++++++++++++++++++++Route IPv4 info+++++++++++++++++++++++++++++ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +-------+-------------+------------+---------------+-----------+-------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | Route | Destination | Gateway | Genmask | Interface | Flags | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +-------+-------------+------------+---------------+-----------+-------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | 0 | 0.0.0.0 | 10.31.12.1 | 0.0.0.0 | eth0 | UG | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | 1 | 10.31.12.0 | 0.0.0.0 | 255.255.252.0 | eth0 | U | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +-------+-------------+------------+---------------+-----------+-------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +++++++++++++++++++Route IPv6 info+++++++++++++++++++ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +-------+-------------+---------+-----------+-------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | Route | Destination | Gateway | Interface | Flags | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +-------+-------------+---------+-----------+-------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | 0 | fe80::/64 | :: | eth0 | U | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | 2 | multicast | :: | eth0 | U | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +-------+-------------+---------+-----------+-------+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Generating public/private rsa key pair. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Your identification has been saved in /etc/ssh/ssh_host_rsa_key Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Your public key has been saved in /etc/ssh/ssh_host_rsa_key.pub Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: The key fingerprint is: Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: SHA256:6zYh8D9sJVTMc/UplqhDb1JRoJSmKvg9yVrtJKOf8Y0 root@ip-10-31-14-110.us-east-1.aws.redhat.com Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: The key's randomart image is: Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: +---[RSA 3072]----+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | +.ooo. | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | .o* + ...| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | o+ = + ..| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . .o + . . | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . o..S+ o | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . . .+ o.= | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . +=.*.+ | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | .o=O.X | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | ooo.E.+ | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: +----[SHA256]-----+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Generating public/private ecdsa key pair. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Your identification has been saved in /etc/ssh/ssh_host_ecdsa_key Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Your public key has been saved in /etc/ssh/ssh_host_ecdsa_key.pub Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: The key fingerprint is: Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: SHA256:2/VGYFFBaPbV9mIKV6Gxe7Knrqzyong5LsvYtXW93UY root@ip-10-31-14-110.us-east-1.aws.redhat.com Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: The key's randomart image is: Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: +---[ECDSA 256]---+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | o+=o.| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | +=. +| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | o=o o.| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | ...o+ .| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | S o+oo. | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | + ..*E | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | ... o o ..+ | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | +.o+oo. . o =. | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: |. ==+o +o.+o+.. | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: +----[SHA256]-----+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Generating public/private ed25519 key pair. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Your identification has been saved in /etc/ssh/ssh_host_ed25519_key Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Your public key has been saved in /etc/ssh/ssh_host_ed25519_key.pub Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: The key fingerprint is: Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: SHA256:nR3w4spyXF/iKNEXvkwO+019RpKTGbhfuvuA8uYmPwU root@ip-10-31-14-110.us-east-1.aws.redhat.com Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: The key's randomart image is: Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: +--[ED25519 256]--+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | o . | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . = . | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | + =E+ = | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | S B B.B o| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | o + @ *.B | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . * + B.= +| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | o ..++o +.| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | *=.+o.| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: +----[SHA256]-----+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished cloud-init.service - Initial cloud-init job (metadata service crawler). ░░ Subject: A start job for unit cloud-init.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.service has finished successfully. ░░ ░░ The job identifier is 267. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target cloud-config.target - Cloud-config availability. ░░ Subject: A start job for unit cloud-config.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.target has finished successfully. ░░ ░░ The job identifier is 266. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target network-online.target - Network is Online. ░░ Subject: A start job for unit network-online.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network-online.target has finished successfully. ░░ ░░ The job identifier is 205. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting cloud-config.service - Apply the settings specified in cloud-config... ░░ Subject: A start job for unit cloud-config.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.service has begun execution. ░░ ░░ The job identifier is 265. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting kdump.service - Crash recovery kernel arming... ░░ Subject: A start job for unit kdump.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit kdump.service has begun execution. ░░ ░░ The job identifier is 237. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting restraintd.service - The restraint harness.... ░░ Subject: A start job for unit restraintd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit restraintd.service has begun execution. ░░ ░░ The job identifier is 278. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting rpc-statd-notify.service - Notify NFS peers of a restart... ░░ Subject: A start job for unit rpc-statd-notify.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-statd-notify.service has begun execution. ░░ ░░ The job identifier is 246. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting sshd.service - OpenSSH server daemon... ░░ Subject: A start job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 269. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com sm-notify[876]: Version 2.7.1 starting Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started restraintd.service - The restraint harness.. ░░ Subject: A start job for unit restraintd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit restraintd.service has finished successfully. ░░ ░░ The job identifier is 278. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com (sshd)[879]: sshd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started rpc-statd-notify.service - Notify NFS peers of a restart. ░░ Subject: A start job for unit rpc-statd-notify.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-statd-notify.service has finished successfully. ░░ ░░ The job identifier is 246. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com sshd[879]: Server listening on 0.0.0.0 port 22. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com sshd[879]: Server listening on :: port 22. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started sshd.service - OpenSSH server daemon. ░░ Subject: A start job for unit sshd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has finished successfully. ░░ ░░ The job identifier is 269. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[920]: Cloud-init v. 24.1.4-21.el10 running 'modules:config' at Sat, 14 Dec 2024 16:26:41 +0000. Up 22.75 seconds. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Stopping sshd.service - OpenSSH server daemon... ░░ Subject: A stop job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 501. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com sshd[879]: Received signal 15; terminating. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: sshd.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit sshd.service has successfully entered the 'dead' state. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Stopped sshd.service - OpenSSH server daemon. ░░ Subject: A stop job for unit sshd.service has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd.service has finished. ░░ ░░ The job identifier is 501 and the job result is done. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Stopped target sshd-keygen.target. ░░ Subject: A stop job for unit sshd-keygen.target has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd-keygen.target has finished. ░░ ░░ The job identifier is 585 and the job result is done. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Stopping sshd-keygen.target... ░░ Subject: A stop job for unit sshd-keygen.target has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd-keygen.target has begun execution. ░░ ░░ The job identifier is 585. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: ssh-host-keys-migration.service - Update OpenSSH host key permissions was skipped because of an unmet condition check (ConditionPathExists=!/var/lib/.ssh-host-keys-migration). ░░ Subject: A start job for unit ssh-host-keys-migration.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit ssh-host-keys-migration.service has finished successfully. ░░ ░░ The job identifier is 584. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: sshd-keygen@ecdsa.service - OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ecdsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ecdsa.service has finished successfully. ░░ ░░ The job identifier is 583. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: sshd-keygen@ed25519.service - OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ed25519.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ed25519.service has finished successfully. ░░ ░░ The job identifier is 582. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: sshd-keygen@rsa.service - OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@rsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@rsa.service has finished successfully. ░░ ░░ The job identifier is 580. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target sshd-keygen.target. ░░ Subject: A start job for unit sshd-keygen.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen.target has finished successfully. ░░ ░░ The job identifier is 585. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting sshd.service - OpenSSH server daemon... ░░ Subject: A start job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 501. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com (sshd)[924]: sshd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com sshd[924]: Server listening on 0.0.0.0 port 22. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com sshd[924]: Server listening on :: port 22. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started sshd.service - OpenSSH server daemon. ░░ Subject: A start job for unit sshd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has finished successfully. ░░ ░░ The job identifier is 501. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished cloud-config.service - Apply the settings specified in cloud-config. ░░ Subject: A start job for unit cloud-config.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.service has finished successfully. ░░ ░░ The job identifier is 265. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting cloud-final.service - Execute cloud user/final scripts... ░░ Subject: A start job for unit cloud-final.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-final.service has begun execution. ░░ ░░ The job identifier is 276. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... ░░ Subject: A start job for unit systemd-user-sessions.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-user-sessions.service has begun execution. ░░ ░░ The job identifier is 262. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. ░░ Subject: A start job for unit systemd-user-sessions.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-user-sessions.service has finished successfully. ░░ ░░ The job identifier is 262. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started crond.service - Command Scheduler. ░░ Subject: A start job for unit crond.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit crond.service has finished successfully. ░░ ░░ The job identifier is 260. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started getty@tty1.service - Getty on tty1. ░░ Subject: A start job for unit getty@tty1.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit getty@tty1.service has finished successfully. ░░ ░░ The job identifier is 232. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. ░░ Subject: A start job for unit serial-getty@ttyS0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit serial-getty@ttyS0.service has finished successfully. ░░ ░░ The job identifier is 227. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target getty.target - Login Prompts. ░░ Subject: A start job for unit getty.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit getty.target has finished successfully. ░░ ░░ The job identifier is 226. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target multi-user.target - Multi-User System. ░░ Subject: A start job for unit multi-user.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit multi-user.target has finished successfully. ░░ ░░ The job identifier is 121. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting systemd-update-utmp-runlevel.service - Record Runlevel Change in UTMP... ░░ Subject: A start job for unit systemd-update-utmp-runlevel.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp-runlevel.service has begun execution. ░░ ░░ The job identifier is 235. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com restraintd[883]: Listening on http://localhost:8081 Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com crond[930]: (CRON) STARTUP (1.7.0) Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com crond[930]: (CRON) INFO (Syslog will be used instead of sendmail.) Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com crond[930]: (CRON) INFO (RANDOM_DELAY will be scaled with factor 7% if used.) Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com crond[930]: (CRON) INFO (running with inotify support) Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-update-utmp-runlevel.service has successfully entered the 'dead' state. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished systemd-update-utmp-runlevel.service - Record Runlevel Change in UTMP. ░░ Subject: A start job for unit systemd-update-utmp-runlevel.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp-runlevel.service has finished successfully. ░░ ░░ The job identifier is 235. Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com kdumpctl[881]: kdump: Detected change(s) in the following file(s): /etc/fstab Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1035]: Cloud-init v. 24.1.4-21.el10 running 'modules:final' at Sat, 14 Dec 2024 16:26:42 +0000. Up 23.23 seconds. Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1080]: ############################################################# Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1081]: -----BEGIN SSH HOST KEY FINGERPRINTS----- Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1083]: 256 SHA256:2/VGYFFBaPbV9mIKV6Gxe7Knrqzyong5LsvYtXW93UY root@ip-10-31-14-110.us-east-1.aws.redhat.com (ECDSA) Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1088]: 256 SHA256:nR3w4spyXF/iKNEXvkwO+019RpKTGbhfuvuA8uYmPwU root@ip-10-31-14-110.us-east-1.aws.redhat.com (ED25519) Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1094]: 3072 SHA256:6zYh8D9sJVTMc/UplqhDb1JRoJSmKvg9yVrtJKOf8Y0 root@ip-10-31-14-110.us-east-1.aws.redhat.com (RSA) Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1098]: -----END SSH HOST KEY FINGERPRINTS----- Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1099]: ############################################################# Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1035]: Cloud-init v. 24.1.4-21.el10 finished at Sat, 14 Dec 2024 16:26:42 +0000. Datasource DataSourceEc2Local. Up 23.40 seconds Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished cloud-final.service - Execute cloud user/final scripts. ░░ Subject: A start job for unit cloud-final.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-final.service has finished successfully. ░░ ░░ The job identifier is 276. Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target cloud-init.target - Cloud-init target. ░░ Subject: A start job for unit cloud-init.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.target has finished successfully. ░░ ░░ The job identifier is 264. Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 0 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 0 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 48 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 48 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 49 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 49 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 50 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 50 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 51 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 51 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 52 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 52 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 53 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 53 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 54 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 54 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 55 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 55 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 56 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 56 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 57 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 57 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 58 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 58 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 59 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 59 affinity is now unmanaged Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com kernel: block xvda: the capability attribute has been deprecated. Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com kdumpctl[881]: kdump: Rebuilding /boot/initramfs-6.12.0-31.el10.x86_64kdump.img Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1388]: dracut-103-1.el10 Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1391]: Executing: /usr/bin/dracut --list-modules Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1462]: dracut-103-1.el10 Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Executing: /usr/bin/dracut --add kdumpbase --quiet --hostonly --hostonly-cmdline --hostonly-i18n --hostonly-mode strict --hostonly-nics --aggressive-strip --omit "rdma plymouth resume ifcfg earlykdump" --mount "/dev/disk/by-uuid/f3bb1e80-fac3-4b5e-93f6-d763469176c6 /sysroot xfs rw,relatime,seclabel,attr2,inode64,logbufs=8,logbsize=32k,noquota" --add squash-squashfs --squash-compressor zstd --no-hostonly-default-device -f /boot/initramfs-6.12.0-31.el10.x86_64kdump.img 6.12.0-31.el10.x86_64 Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-bsod' will not be installed, because command '/usr/lib/systemd/systemd-bsod' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-networkd' will not be installed, because command '/usr/lib/systemd/systemd-networkd' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-networkd' will not be installed, because command '/usr/lib/systemd/systemd-networkd-wait-online' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-pcrphase' will not be installed, because command '/usr/lib/systemd/systemd-pcrphase' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-portabled' will not be installed, because command 'portablectl' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-portabled' will not be installed, because command '/usr/lib/systemd/systemd-portabled' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-resolved' will not be installed, because command '/usr/lib/systemd/systemd-resolved' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-timesyncd' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-time-wait-sync' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'busybox' will not be installed, because command 'busybox' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com chronyd[662]: Selected source 10.2.32.38 Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'dbus-daemon' will not be installed, because command 'dbus-daemon' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'connman' will not be installed, because command 'connmand' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'connman' will not be installed, because command 'connmanctl' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'connman' will not be installed, because command 'connmand-wait-online' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'ifcfg' will not be installed, because it's in the list to be omitted! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'plymouth' will not be installed, because it's in the list to be omitted! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: 62bluetooth: Could not find any command of '/usr/lib/bluetooth/bluetoothd /usr/libexec/bluetooth/bluetoothd'! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'btrfs' will not be installed, because command 'btrfs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'dmraid' will not be installed, because command 'dmraid' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'mdraid' will not be installed, because command 'mdadm' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'multipath' will not be installed, because command 'multipath' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'iscsi' will not be installed, because command 'iscsid' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'nvmf' will not be installed, because command 'nvme' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'resume' will not be installed, because it's in the list to be omitted! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'squash-erofs' will not be installed, because command 'mkfs.erofs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'squash-erofs' will not be installed, because command 'fsck.erofs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'biosdevname' will not be installed, because command 'biosdevname' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'earlykdump' will not be installed, because it's in the list to be omitted! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-bsod' will not be installed, because command '/usr/lib/systemd/systemd-bsod' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-pcrphase' will not be installed, because command '/usr/lib/systemd/systemd-pcrphase' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-portabled' will not be installed, because command 'portablectl' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-portabled' will not be installed, because command '/usr/lib/systemd/systemd-portabled' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-resolved' will not be installed, because command '/usr/lib/systemd/systemd-resolved' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-timesyncd' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-time-wait-sync' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'busybox' will not be installed, because command 'busybox' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'dbus-daemon' will not be installed, because command 'dbus-daemon' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'connman' will not be installed, because command 'connmand' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'connman' will not be installed, because command 'connmanctl' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'connman' will not be installed, because command 'connmand-wait-online' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: 62bluetooth: Could not find any command of '/usr/lib/bluetooth/bluetoothd /usr/libexec/bluetooth/bluetoothd'! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'btrfs' will not be installed, because command 'btrfs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'dmraid' will not be installed, because command 'dmraid' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'mdraid' will not be installed, because command 'mdadm' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'multipath' will not be installed, because command 'multipath' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'iscsi' will not be installed, because command 'iscsid' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'nvmf' will not be installed, because command 'nvme' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'squash-erofs' will not be installed, because command 'mkfs.erofs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'squash-erofs' will not be installed, because command 'fsck.erofs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: fips *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: fips-crypto-policies *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-ask-password *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-initrd *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-journald *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-modules-load *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-sysctl *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-sysusers *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-tmpfiles *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-udevd *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: rngd *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: i18n *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: drm *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: prefixdevname *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: kernel-modules *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: kernel-modules-extra *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: kernel-modules-extra: configuration source "/run/depmod.d" does not exist Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: kernel-modules-extra: configuration source "/lib/depmod.d" does not exist Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: kernel-modules-extra: parsing configuration file "/etc/depmod.d/dist.conf" Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: kernel-modules-extra: /etc/depmod.d/dist.conf: added "updates extra built-in weak-updates" to the list of search directories Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: pcmcia *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Skipping udev rule: 60-pcmcia.rules Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: fstab-sys *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: hwdb *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: rootfs-block *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: squash-squashfs *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: terminfo *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: udev-rules *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: dracut-systemd *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: usrmount *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: base *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: fs-lib *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: kdumpbase *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: memstrack *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: microcode_ctl-fw_dir_override *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl module: mangling fw_dir Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl: reset fw_dir to "/lib/firmware/updates /lib/firmware" Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel"... Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl: intel: caveats check for kernel version "6.12.0-31.el10.x86_64" passed, adding "/usr/share/microcode_ctl/ucode_with_caveats/intel" to fw_dir variable Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-4f-01"... Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl: configuration "intel-06-4f-01" is ignored Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl: final fw_dir: "/usr/share/microcode_ctl/ucode_with_caveats/intel /lib/firmware/updates /lib/firmware" Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: shutdown *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: squash-lib *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including modules done *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Installing kernel module dependencies *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Installing kernel module dependencies done *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Resolving executable dependencies *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Resolving executable dependencies done *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Hardlinking files *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Mode: real Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Method: sha256 Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Files: 537 Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Linked: 25 files Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Compared: 0 xattrs Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Compared: 48 files Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Saved: 13.58 MiB Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Duration: 0.152061 seconds Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Hardlinking files done *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Generating early-microcode cpio image *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Constructing GenuineIntel.bin *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Constructing GenuineIntel.bin *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Store current command line parameters *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Stored kernel commandline: Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: No dracut internal kernel commandline stored in the initramfs Dec 14 11:26:51 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Squashing the files inside the initramfs *** Dec 14 11:26:58 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Squashing the files inside the initramfs done *** Dec 14 11:26:58 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Creating image file '/boot/initramfs-6.12.0-31.el10.x86_64kdump.img' *** Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Creating initramfs image file '/boot/initramfs-6.12.0-31.el10.x86_64kdump.img' done *** Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com kernel: PKCS7: Message signed outside of X.509 validity window Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com kdumpctl[881]: kdump: kexec: loaded kdump kernel Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com kdumpctl[881]: kdump: Starting kdump: [OK] Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com kdumpctl[881]: kdump: Notice: No vmcore creation test performed! Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished kdump.service - Crash recovery kernel arming. ░░ Subject: A start job for unit kdump.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit kdump.service has finished successfully. ░░ ░░ The job identifier is 237. Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Startup finished in 995ms (kernel) + 5.404s (initrd) + 34.706s (userspace) = 41.106s. ░░ Subject: System start-up is now complete ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ All system services necessary queued for starting at boot have been ░░ started. Note that this does not mean that the machine is now idle as services ░░ might still be busy with completing start-up. ░░ ░░ Kernel start-up required 995066 microseconds. ░░ ░░ Initrd start-up required 5404966 microseconds. ░░ ░░ Userspace start-up required 34706427 microseconds. Dec 14 11:27:09 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: systemd-hostnamed.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-hostnamed.service has successfully entered the 'dead' state. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4326]: Accepted publickey for root from 10.30.34.106 port 56984 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4326]: pam_systemd(sshd:session): New sd-bus connection (system-bus-pam-systemd-4326) opened. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Created slice user-0.slice - User Slice of UID 0. ░░ Subject: A start job for unit user-0.slice has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-0.slice has finished successfully. ░░ ░░ The job identifier is 673. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting user-runtime-dir@0.service - User Runtime Directory /run/user/0... ░░ Subject: A start job for unit user-runtime-dir@0.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-runtime-dir@0.service has begun execution. ░░ ░░ The job identifier is 595. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: New session 1 of user root. ░░ Subject: A new session 1 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 1 has been created for the user root. ░░ ░░ The leading process of the session is 4326. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished user-runtime-dir@0.service - User Runtime Directory /run/user/0. ░░ Subject: A start job for unit user-runtime-dir@0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-runtime-dir@0.service has finished successfully. ░░ ░░ The job identifier is 595. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting user@0.service - User Manager for UID 0... ░░ Subject: A start job for unit user@0.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user@0.service has begun execution. ░░ ░░ The job identifier is 675. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: New session 2 of user root. ░░ Subject: A new session 2 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 2 has been created for the user root. ░░ ░░ The leading process of the session is 4331. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com (systemd)[4331]: pam_unix(systemd-user:session): session opened for user root(uid=0) by root(uid=0) Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Queued start job for default target default.target. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Created slice app.slice - User Application Slice. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 5. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: grub-boot-success.timer - Mark boot as successful after the user session has run 2 minutes was skipped because of an unmet condition check (ConditionUser=!@system). ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 11. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of User's Temporary Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 10. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Reached target paths.target - Paths. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 8. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Reached target timers.target - Timers. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 9. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Starting dbus.socket - D-Bus User Message Bus Socket... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 4. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Starting systemd-tmpfiles-setup.service - Create User Files and Directories... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 12. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Finished systemd-tmpfiles-setup.service - Create User Files and Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 12. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Listening on dbus.socket - D-Bus User Message Bus Socket. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 4. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Reached target sockets.target - Sockets. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 3. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Reached target basic.target - Basic System. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 2. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Reached target default.target - Main User Target. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 1. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Startup finished in 112ms. ░░ Subject: User manager start-up is now complete ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The user manager instance for user 0 has been started. All services queued ░░ for starting have been started. Note that other services might still be starting ░░ up or be started at any later time. ░░ ░░ Startup of the manager took 112474 microseconds. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started user@0.service - User Manager for UID 0. ░░ Subject: A start job for unit user@0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user@0.service has finished successfully. ░░ ░░ The job identifier is 675. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started session-1.scope - Session 1 of User root. ░░ Subject: A start job for unit session-1.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-1.scope has finished successfully. ░░ ░░ The job identifier is 756. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4326]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4342]: Received disconnect from 10.30.34.106 port 56984:11: disconnected by user Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4342]: Disconnected from user root 10.30.34.106 port 56984 Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4326]: pam_systemd(sshd:session): New sd-bus connection (system-bus-pam-systemd-4326) opened. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4326]: pam_unix(sshd:session): session closed for user root Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: session-1.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-1.scope has successfully entered the 'dead' state. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: Session 1 logged out. Waiting for processes to exit. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: Removed session 1. ░░ Subject: Session 1 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 1 has been terminated. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4380]: Accepted publickey for root from 10.31.8.152 port 46578 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4381]: Accepted publickey for root from 10.31.8.152 port 46592 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4380]: pam_systemd(sshd:session): New sd-bus connection (system-bus-pam-systemd-4380) opened. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4381]: pam_systemd(sshd:session): New sd-bus connection (system-bus-pam-systemd-4381) opened. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: New session 3 of user root. ░░ Subject: A new session 3 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 3 has been created for the user root. ░░ ░░ The leading process of the session is 4380. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started session-3.scope - Session 3 of User root. ░░ Subject: A start job for unit session-3.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-3.scope has finished successfully. ░░ ░░ The job identifier is 838. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: New session 4 of user root. ░░ Subject: A new session 4 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 4 has been created for the user root. ░░ ░░ The leading process of the session is 4381. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started session-4.scope - Session 4 of User root. ░░ Subject: A start job for unit session-4.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-4.scope has finished successfully. ░░ ░░ The job identifier is 920. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4380]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4381]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4387]: Received disconnect from 10.31.8.152 port 46592:11: disconnected by user Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4387]: Disconnected from user root 10.31.8.152 port 46592 Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4381]: pam_systemd(sshd:session): New sd-bus connection (system-bus-pam-systemd-4381) opened. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4381]: pam_unix(sshd:session): session closed for user root Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: Session 4 logged out. Waiting for processes to exit. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: session-4.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-4.scope has successfully entered the 'dead' state. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: Removed session 4. ░░ Subject: Session 4 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 4 has been terminated. Dec 14 11:29:28 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting systemd-hostnamed.service - Hostname Service... ░░ Subject: A start job for unit systemd-hostnamed.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has begun execution. ░░ ░░ The job identifier is 1002. Dec 14 11:29:28 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started systemd-hostnamed.service - Hostname Service. ░░ Subject: A start job for unit systemd-hostnamed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has finished successfully. ░░ ░░ The job identifier is 1002. Dec 14 11:29:28 managed-node3 systemd-hostnamed[5861]: Hostname set to (static) Dec 14 11:29:28 managed-node3 NetworkManager[703]: [1734193768.6417] hostname: static hostname changed from "ip-10-31-14-110.us-east-1.aws.redhat.com" to "managed-node3" Dec 14 11:29:28 managed-node3 systemd[1]: Starting NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 1080. Dec 14 11:29:28 managed-node3 systemd[1]: Started NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 1080. Dec 14 11:29:38 managed-node3 systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Dec 14 11:29:58 managed-node3 systemd[1]: systemd-hostnamed.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-hostnamed.service has successfully entered the 'dead' state. Dec 14 11:30:08 managed-node3 sshd-session[6527]: Accepted publickey for root from 10.31.13.174 port 49050 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Dec 14 11:30:08 managed-node3 sshd-session[6527]: pam_systemd(sshd:session): New sd-bus connection (system-bus-pam-systemd-6527) opened. Dec 14 11:30:08 managed-node3 systemd-logind[655]: New session 5 of user root. ░░ Subject: A new session 5 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 5 has been created for the user root. ░░ ░░ The leading process of the session is 6527. Dec 14 11:30:08 managed-node3 systemd[1]: Started session-5.scope - Session 5 of User root. ░░ Subject: A start job for unit session-5.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-5.scope has finished successfully. ░░ ░░ The job identifier is 1159. Dec 14 11:30:08 managed-node3 sshd-session[6527]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Dec 14 11:30:10 managed-node3 python3.12[6683]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Dec 14 11:30:11 managed-node3 python3.12[6843]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:30:11 managed-node3 python3.12[6974]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:30:13 managed-node3 sudo[7236]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-xclcopowrlvzimvlwodbfevotsixbavv ; /usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1734193813.0266075-7127-254553333473826/AnsiballZ_dnf.py' Dec 14 11:30:13 managed-node3 sudo[7236]: pam_systemd(sudo:session): New sd-bus connection (system-bus-pam-systemd-7236) opened. Dec 14 11:30:13 managed-node3 sudo[7236]: pam_unix(sudo:session): session opened for user root(uid=0) by root(uid=0) Dec 14 11:30:13 managed-node3 python3.12[7239]: ansible-ansible.legacy.dnf Invoked with name=['iptables-nft', 'podman', 'shadow-utils-subid'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Dec 14 11:30:27 managed-node3 kernel: SELinux: Converting 384 SID table entries... Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability network_peer_controls=1 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability open_perms=1 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability extended_socket_class=1 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability always_check_network=0 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability userspace_initial_context=0 Dec 14 11:30:35 managed-node3 kernel: SELinux: Converting 385 SID table entries... Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability network_peer_controls=1 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability open_perms=1 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability extended_socket_class=1 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability always_check_network=0 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability userspace_initial_context=0 Dec 14 11:30:43 managed-node3 kernel: SELinux: Converting 385 SID table entries... Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability network_peer_controls=1 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability open_perms=1 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability extended_socket_class=1 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability always_check_network=0 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability userspace_initial_context=0 Dec 14 11:30:44 managed-node3 setsebool[7313]: The virt_use_nfs policy boolean was changed to 1 by root Dec 14 11:30:44 managed-node3 setsebool[7313]: The virt_sandbox_use_all_caps policy boolean was changed to 1 by root Dec 14 11:30:54 managed-node3 kernel: SELinux: Converting 388 SID table entries... Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability network_peer_controls=1 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability open_perms=1 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability extended_socket_class=1 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability always_check_network=0 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability userspace_initial_context=0 Dec 14 11:31:02 managed-node3 kernel: SELinux: Converting 388 SID table entries... Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability network_peer_controls=1 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability open_perms=1 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability extended_socket_class=1 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability always_check_network=0 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability userspace_initial_context=0 Dec 14 11:31:19 managed-node3 systemd[1]: Started run-r716e48b3056747a5bf3b0cc91f601af1.service - /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r716e48b3056747a5bf3b0cc91f601af1.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r716e48b3056747a5bf3b0cc91f601af1.service has finished successfully. ░░ ░░ The job identifier is 1241. Dec 14 11:31:19 managed-node3 systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1319. Dec 14 11:31:19 managed-node3 systemd[1]: Reload requested from client PID 8038 ('systemctl') (unit session-5.scope)... Dec 14 11:31:19 managed-node3 systemd[1]: Reloading... Dec 14 11:31:19 managed-node3 systemd[1]: Reloading finished in 254 ms. Dec 14 11:31:19 managed-node3 systemd[1]: Queuing reload/restart jobs for marked units… Dec 14 11:31:20 managed-node3 sudo[7236]: pam_unix(sudo:session): session closed for user root Dec 14 11:31:21 managed-node3 python3.12[8493]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:21 managed-node3 systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Dec 14 11:31:21 managed-node3 systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1319. Dec 14 11:31:21 managed-node3 systemd[1]: run-r716e48b3056747a5bf3b0cc91f601af1.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-r716e48b3056747a5bf3b0cc91f601af1.service has successfully entered the 'dead' state. Dec 14 11:31:22 managed-node3 python3.12[8634]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Dec 14 11:31:22 managed-node3 python3.12[8766]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:24 managed-node3 python3.12[8899]: ansible-tempfile Invoked with state=directory prefix=lsr_podman_config_ suffix= path=None Dec 14 11:31:24 managed-node3 python3.12[9030]: ansible-ansible.legacy.command Invoked with _raw_params=tar --ignore-failed-read -c -P -v -p -f /tmp/lsr_podman_config_umo1wg40/backup.tar /etc/containers/containers.conf.d/50-systemroles.conf /etc/containers/registries.conf.d/50-systemroles.conf /etc/containers/storage.conf /etc/containers/policy.json _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:25 managed-node3 python3.12[9162]: ansible-user Invoked with name=user1 state=present non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on managed-node3 update_password=always uid=None group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None password_expire_warn=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None Dec 14 11:31:25 managed-node3 useradd[9164]: new group: name=user1, GID=1000 Dec 14 11:31:25 managed-node3 useradd[9164]: new user: name=user1, UID=1000, GID=1000, home=/home/user1, shell=/bin/bash, from=/dev/pts/0 Dec 14 11:31:27 managed-node3 python3.12[9426]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:28 managed-node3 python3.12[9564]: ansible-getent Invoked with database=passwd key=user1 fail_key=False service=None split=None Dec 14 11:31:29 managed-node3 python3.12[9696]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:29 managed-node3 python3.12[9829]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:29 managed-node3 python3.12[9961]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:30 managed-node3 python3.12[10093]: ansible-file Invoked with path=/home/user1/.config/containers/containers.conf.d state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:31 managed-node3 python3.12[10224]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:31 managed-node3 python3.12[10329]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734193891.0325742-8177-94420310335864/.source.conf dest=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf owner=user1 mode=0644 follow=False _original_basename=toml.j2 checksum=94370d6e765779f1c58daf02f667b8f0b74d91f6 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:32 managed-node3 python3.12[10460]: ansible-file Invoked with path=/home/user1/.config/containers/registries.conf.d state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:32 managed-node3 python3.12[10591]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:32 managed-node3 python3.12[10696]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734193892.3603973-8232-1526005629345/.source.conf dest=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf owner=user1 mode=0644 follow=False _original_basename=toml.j2 checksum=dfb9cd7094a81b3d1bb06512cc9b49a09c75639b backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:33 managed-node3 python3.12[10827]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:34 managed-node3 python3.12[10958]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:34 managed-node3 python3.12[11063]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734193893.732546-8296-55796330500548/.source.conf dest=/home/user1/.config/containers/storage.conf owner=user1 mode=0644 follow=False _original_basename=toml.j2 checksum=d08574b6a1df63dbe1c939ff0bcc7c0b61d03044 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:34 managed-node3 python3.12[11194]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:35 managed-node3 python3.12[11325]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:35 managed-node3 python3.12[11456]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:36 managed-node3 python3.12[11561]: ansible-ansible.legacy.copy Invoked with dest=/home/user1/.config/containers/policy.json owner=user1 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1734193895.4946458-8411-208121650010930/.source.json _original_basename=.tddnhefz follow=False checksum=6746c079ad563b735fc39f73d4876654b80b0a0d backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:36 managed-node3 python3.12[11692]: ansible-stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:37 managed-node3 python3.12[11825]: ansible-stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:37 managed-node3 python3.12[11958]: ansible-stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:37 managed-node3 python3.12[12091]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:39 managed-node3 python3.12[12355]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:40 managed-node3 python3.12[12492]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:41 managed-node3 python3.12[12625]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:41 managed-node3 python3.12[12757]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:42 managed-node3 python3.12[12889]: ansible-file Invoked with path=/home/user1/.config/containers/containers.conf.d state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:42 managed-node3 python3.12[13020]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:42 managed-node3 python3.12[13086]: ansible-ansible.legacy.file Invoked with owner=user1 mode=0644 dest=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf _original_basename=toml.j2 recurse=False state=file path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:43 managed-node3 python3.12[13217]: ansible-file Invoked with path=/home/user1/.config/containers/registries.conf.d state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:43 managed-node3 python3.12[13348]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:44 managed-node3 python3.12[13414]: ansible-ansible.legacy.file Invoked with owner=user1 mode=0644 dest=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf _original_basename=toml.j2 recurse=False state=file path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:44 managed-node3 python3.12[13545]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:45 managed-node3 python3.12[13676]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:45 managed-node3 python3.12[13742]: ansible-ansible.legacy.file Invoked with owner=user1 mode=0644 dest=/home/user1/.config/containers/storage.conf _original_basename=toml.j2 recurse=False state=file path=/home/user1/.config/containers/storage.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:45 managed-node3 python3.12[13873]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:46 managed-node3 python3.12[14004]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:46 managed-node3 python3.12[14137]: ansible-slurp Invoked with path=/home/user1/.config/containers/policy.json src=/home/user1/.config/containers/policy.json Dec 14 11:31:47 managed-node3 python3.12[14268]: ansible-stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:48 managed-node3 python3.12[14401]: ansible-stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:48 managed-node3 python3.12[14534]: ansible-stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:48 managed-node3 python3.12[14667]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:51 managed-node3 python3.12[14931]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:52 managed-node3 python3.12[15069]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Dec 14 11:31:52 managed-node3 python3.12[15201]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:54 managed-node3 python3.12[15334]: ansible-file Invoked with path=/etc/containers/containers.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:54 managed-node3 python3.12[15465]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:55 managed-node3 python3.12[15570]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734193914.33953-9204-179873555081338/.source.conf dest=/etc/containers/containers.conf.d/50-systemroles.conf owner=root mode=0644 follow=False _original_basename=toml.j2 checksum=94370d6e765779f1c58daf02f667b8f0b74d91f6 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:55 managed-node3 python3.12[15701]: ansible-file Invoked with path=/etc/containers/registries.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:56 managed-node3 python3.12[15832]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:56 managed-node3 python3.12[15937]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734193915.975161-9276-112706284746257/.source.conf dest=/etc/containers/registries.conf.d/50-systemroles.conf owner=root mode=0644 follow=False _original_basename=toml.j2 checksum=dfb9cd7094a81b3d1bb06512cc9b49a09c75639b backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:57 managed-node3 python3.12[16068]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:57 managed-node3 python3.12[16199]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:58 managed-node3 python3.12[16304]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734193917.5309792-9356-138504811101551/.source.conf dest=/etc/containers/storage.conf owner=root mode=0644 follow=False _original_basename=toml.j2 checksum=d08574b6a1df63dbe1c939ff0bcc7c0b61d03044 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:58 managed-node3 python3.12[16435]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:59 managed-node3 python3.12[16566]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:59 managed-node3 python3.12[16699]: ansible-slurp Invoked with path=/etc/containers/policy.json src=/etc/containers/policy.json Dec 14 11:32:00 managed-node3 python3.12[16830]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:32:00 managed-node3 python3.12[16937]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/policy.json owner=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1734193919.7730489-9479-127190054442142/.source.json _original_basename=.d8o5zfsi follow=False checksum=6746c079ad563b735fc39f73d4876654b80b0a0d backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:01 managed-node3 python3.12[17068]: ansible-stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:02 managed-node3 python3.12[17201]: ansible-stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:02 managed-node3 python3.12[17334]: ansible-stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:02 managed-node3 python3.12[17467]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:05 managed-node3 python3.12[17731]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:32:06 managed-node3 python3.12[17868]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:07 managed-node3 python3.12[18001]: ansible-file Invoked with path=/etc/containers/containers.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:08 managed-node3 python3.12[18132]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:32:08 managed-node3 python3.12[18198]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/containers.conf.d/50-systemroles.conf _original_basename=toml.j2 recurse=False state=file path=/etc/containers/containers.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:09 managed-node3 python3.12[18329]: ansible-file Invoked with path=/etc/containers/registries.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:09 managed-node3 python3.12[18460]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:32:09 managed-node3 python3.12[18526]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/registries.conf.d/50-systemroles.conf _original_basename=toml.j2 recurse=False state=file path=/etc/containers/registries.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:10 managed-node3 python3.12[18657]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:11 managed-node3 python3.12[18788]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:32:11 managed-node3 python3.12[18854]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/storage.conf _original_basename=toml.j2 recurse=False state=file path=/etc/containers/storage.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:11 managed-node3 python3.12[18985]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:12 managed-node3 python3.12[19116]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:12 managed-node3 python3.12[19249]: ansible-slurp Invoked with path=/etc/containers/policy.json src=/etc/containers/policy.json Dec 14 11:32:14 managed-node3 python3.12[19380]: ansible-stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:14 managed-node3 python3.12[19513]: ansible-stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:14 managed-node3 python3.12[19646]: ansible-stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:15 managed-node3 python3.12[19779]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:16 managed-node3 python3.12[19912]: ansible-slurp Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf src=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf Dec 14 11:32:16 managed-node3 python3.12[20043]: ansible-slurp Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf src=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf Dec 14 11:32:17 managed-node3 python3.12[20174]: ansible-slurp Invoked with path=/home/user1/.config/containers/storage.conf src=/home/user1/.config/containers/storage.conf Dec 14 11:32:17 managed-node3 python3.12[20305]: ansible-slurp Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf src=/etc/containers/containers.conf.d/50-systemroles.conf Dec 14 11:32:18 managed-node3 python3.12[20436]: ansible-slurp Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf src=/etc/containers/registries.conf.d/50-systemroles.conf Dec 14 11:32:19 managed-node3 python3.12[20567]: ansible-slurp Invoked with path=/etc/containers/storage.conf src=/etc/containers/storage.conf Dec 14 11:32:19 managed-node3 python3.12[20698]: ansible-file Invoked with state=absent path=/etc/containers/containers.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:20 managed-node3 python3.12[20829]: ansible-file Invoked with state=absent path=/etc/containers/registries.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:20 managed-node3 python3.12[20960]: ansible-file Invoked with state=absent path=/etc/containers/storage.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:21 managed-node3 python3.12[21091]: ansible-file Invoked with state=absent path=/etc/containers/policy.json recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:21 managed-node3 python3.12[21222]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:21 managed-node3 python3.12[21353]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:22 managed-node3 python3.12[21484]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/storage.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:22 managed-node3 python3.12[21615]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/policy.json recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:23 managed-node3 python3.12[21746]: ansible-ansible.legacy.command Invoked with _raw_params=tar xfvpP /tmp/lsr_podman_config_umo1wg40/backup.tar _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:32:23 managed-node3 python3.12[21878]: ansible-file Invoked with state=absent path=/tmp/lsr_podman_config_umo1wg40 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:26 managed-node3 python3.12[22052]: ansible-setup Invoked with gather_subset=['!all', '!min', 'distribution', 'distribution_major_version', 'distribution_version', 'os_family'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Dec 14 11:32:26 managed-node3 python3.12[22185]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:27 managed-node3 python3.12[22316]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:29 managed-node3 python3.12[22578]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:32:30 managed-node3 python3.12[22715]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Dec 14 11:32:31 managed-node3 python3.12[22847]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:35 managed-node3 python3.12[23023]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Dec 14 11:32:40 managed-node3 python3.12[23183]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:40 managed-node3 python3.12[23314]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:43 managed-node3 python3.12[23576]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:32:44 managed-node3 python3.12[23714]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Dec 14 11:32:45 managed-node3 python3.12[23846]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:51 managed-node3 python3.12[24022]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Dec 14 11:32:53 managed-node3 python3.12[24182]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:53 managed-node3 python3.12[24313]: ansible-ansible.legacy.dnf Invoked with name=['python3-pyasn1', 'python3-cryptography', 'python3-dbus'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Dec 14 11:32:56 managed-node3 python3.12[24449]: ansible-ansible.legacy.dnf Invoked with name=['certmonger', 'python3-packaging'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Dec 14 11:32:57 managed-node3 dbus-broker-launch[646]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Dec 14 11:32:57 managed-node3 dbus-broker-launch[646]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Dec 14 11:32:57 managed-node3 dbus-broker-launch[646]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Dec 14 11:32:57 managed-node3 systemd[1]: Reload requested from client PID 24457 ('systemctl') (unit session-5.scope)... Dec 14 11:32:57 managed-node3 systemd[1]: Reloading... Dec 14 11:32:58 managed-node3 systemd[1]: Reloading finished in 185 ms. Dec 14 11:32:58 managed-node3 systemd[1]: Started run-re25853d457974692b318b9a879e60844.service - /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-re25853d457974692b318b9a879e60844.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-re25853d457974692b318b9a879e60844.service has finished successfully. ░░ ░░ The job identifier is 1401. Dec 14 11:32:58 managed-node3 systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1479. Dec 14 11:32:58 managed-node3 systemd[1]: Reload requested from client PID 24519 ('systemctl') (unit session-5.scope)... Dec 14 11:32:58 managed-node3 systemd[1]: Reloading... Dec 14 11:32:58 managed-node3 systemd[1]: Reloading finished in 291 ms. Dec 14 11:32:58 managed-node3 systemd[1]: Queuing reload/restart jobs for marked units… Dec 14 11:32:59 managed-node3 systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Dec 14 11:32:59 managed-node3 systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1479. Dec 14 11:32:59 managed-node3 systemd[1]: run-re25853d457974692b318b9a879e60844.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-re25853d457974692b318b9a879e60844.service has successfully entered the 'dead' state. Dec 14 11:32:59 managed-node3 python3.12[24715]: ansible-file Invoked with name=/etc/certmonger//pre-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//pre-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:00 managed-node3 python3.12[24846]: ansible-file Invoked with name=/etc/certmonger//post-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//post-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:00 managed-node3 python3.12[24977]: ansible-ansible.legacy.systemd Invoked with name=certmonger state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Dec 14 11:33:00 managed-node3 systemd[1]: Reload requested from client PID 24980 ('systemctl') (unit session-5.scope)... Dec 14 11:33:00 managed-node3 systemd[1]: Reloading... Dec 14 11:33:01 managed-node3 systemd[1]: Reloading finished in 181 ms. Dec 14 11:33:01 managed-node3 systemd[1]: Starting certmonger.service - Certificate monitoring and PKI enrollment... ░░ Subject: A start job for unit certmonger.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit certmonger.service has begun execution. ░░ ░░ The job identifier is 1557. Dec 14 11:33:01 managed-node3 (rtmonger)[25034]: certmonger.service: Referenced but unset environment variable evaluates to an empty string: OPTS Dec 14 11:33:01 managed-node3 systemd[1]: Started certmonger.service - Certificate monitoring and PKI enrollment. ░░ Subject: A start job for unit certmonger.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit certmonger.service has finished successfully. ░░ ░░ The job identifier is 1557. Dec 14 11:33:01 managed-node3 python3.12[25192]: ansible-fedora.linux_system_roles.certificate_request Invoked with name=quadlet_demo dns=['localhost'] directory=/etc/pki/tls wait=True ca=self-sign __header=# # Ansible managed # # system_role:certificate provider_config_directory=/etc/certmonger provider=certmonger key_usage=['digitalSignature', 'keyEncipherment'] extended_key_usage=['id-kp-serverAuth', 'id-kp-clientAuth'] auto_renew=True ip=None email=None common_name=None country=None state=None locality=None organization=None organizational_unit=None contact_email=None key_size=None owner=None group=None mode=None principal=None run_before=None run_after=None Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25207]: Certificate in file "/etc/pki/tls/certs/quadlet_demo.crt" issued by CA and saved. Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 python3.12[25338]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Dec 14 11:33:03 managed-node3 python3.12[25469]: ansible-slurp Invoked with path=/etc/pki/tls/private/quadlet_demo.key src=/etc/pki/tls/private/quadlet_demo.key Dec 14 11:33:03 managed-node3 python3.12[25600]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Dec 14 11:33:04 managed-node3 python3.12[25731]: ansible-ansible.legacy.command Invoked with _raw_params=getcert stop-tracking -f /etc/pki/tls/certs/quadlet_demo.crt _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:33:04 managed-node3 certmonger[25034]: 2024-12-14 11:33:04 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:04 managed-node3 python3.12[25863]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:05 managed-node3 python3.12[25994]: ansible-file Invoked with path=/etc/pki/tls/private/quadlet_demo.key state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:05 managed-node3 python3.12[26125]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:06 managed-node3 python3.12[26256]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:06 managed-node3 python3.12[26387]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:08 managed-node3 python3.12[26649]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:33:10 managed-node3 python3.12[26787]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Dec 14 11:33:10 managed-node3 python3.12[26919]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:13 managed-node3 python3.12[27052]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:14 managed-node3 python3.12[27183]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:14 managed-node3 python3.12[27314]: ansible-ansible.legacy.dnf Invoked with name=['firewalld'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Dec 14 11:33:16 managed-node3 python3.12[27446]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False scope=system no_block=False state=None enabled=None force=None Dec 14 11:33:16 managed-node3 python3.12[27579]: ansible-ansible.legacy.systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Dec 14 11:33:16 managed-node3 systemd[1]: Reload requested from client PID 27582 ('systemctl') (unit session-5.scope)... Dec 14 11:33:16 managed-node3 systemd[1]: Reloading... Dec 14 11:33:16 managed-node3 systemd[1]: Reloading finished in 185 ms. Dec 14 11:33:17 managed-node3 systemd[1]: Starting firewalld.service - firewalld - dynamic firewall daemon... ░░ Subject: A start job for unit firewalld.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit firewalld.service has begun execution. ░░ ░░ The job identifier is 1636. Dec 14 11:33:17 managed-node3 systemd[1]: Started firewalld.service - firewalld - dynamic firewall daemon. ░░ Subject: A start job for unit firewalld.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit firewalld.service has finished successfully. ░░ ░░ The job identifier is 1636. Dec 14 11:33:17 managed-node3 systemd[1]: Starting polkit.service - Authorization Manager... ░░ Subject: A start job for unit polkit.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit polkit.service has begun execution. ░░ ░░ The job identifier is 1720. Dec 14 11:33:17 managed-node3 polkitd[27747]: Started polkitd version 125 Dec 14 11:33:17 managed-node3 systemd[1]: Started polkit.service - Authorization Manager. ░░ Subject: A start job for unit polkit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit polkit.service has finished successfully. ░░ ░░ The job identifier is 1720. Dec 14 11:33:18 managed-node3 python3.12[27820]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Dec 14 11:33:18 managed-node3 python3.12[27951]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Dec 14 11:33:20 managed-node3 systemd[1]: var-lib-containers-storage-overlay-compat279581914-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay-compat279581914-merged.mount has successfully entered the 'dead' state. Dec 14 11:33:20 managed-node3 kernel: evm: overlay not supported Dec 14 11:33:20 managed-node3 podman[28090]: 2024-12-14 11:33:20.380826987 -0500 EST m=+0.086593393 system refresh Dec 14 11:33:21 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:21 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:21 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:22 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:22 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:24 managed-node3 python3.12[28560]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:25 managed-node3 python3.12[28693]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:26 managed-node3 python3.12[28824]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo.network follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:33:26 managed-node3 python3.12[28929]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734194006.0269196-14057-133971458264709/.source.network dest=/etc/containers/systemd/quadlet-demo.network owner=root group=0 mode=0644 _original_basename=quadlet-demo.network follow=False checksum=e57c08d49aff4bae8daab138d913aeddaa8682a0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:27 managed-node3 python3.12[29060]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:33:27 managed-node3 systemd[1]: Reload requested from client PID 29061 ('systemctl') (unit session-5.scope)... Dec 14 11:33:27 managed-node3 systemd[1]: Reloading... Dec 14 11:33:27 managed-node3 systemd[1]: Reloading finished in 189 ms. Dec 14 11:33:28 managed-node3 python3.12[29244]: ansible-systemd Invoked with name=quadlet-demo-network.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Dec 14 11:33:28 managed-node3 systemd[1]: Starting quadlet-demo-network.service... ░░ Subject: A start job for unit quadlet-demo-network.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-network.service has begun execution. ░░ ░░ The job identifier is 1800. Dec 14 11:33:28 managed-node3 quadlet-demo-network[29248]: time="2024-12-14T11:33:28-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" Dec 14 11:33:28 managed-node3 quadlet-demo-network[29248]: systemd-quadlet-demo Dec 14 11:33:28 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:28 managed-node3 systemd[1]: Finished quadlet-demo-network.service. ░░ Subject: A start job for unit quadlet-demo-network.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-network.service has finished successfully. ░░ ░░ The job identifier is 1800. Dec 14 11:33:29 managed-node3 python3.12[29387]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:30 managed-node3 python3.12[29520]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:30 managed-node3 python3.12[29651]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:33:31 managed-node3 python3.12[29756]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734194010.6736438-14236-92540342954756/.source.volume dest=/etc/containers/systemd/quadlet-demo-mysql.volume owner=root group=0 mode=0644 _original_basename=quadlet-demo-mysql.volume follow=False checksum=585f8cbdf0ec73000f9227dcffbef71e9552ea4a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:31 managed-node3 python3.12[29887]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:33:31 managed-node3 systemd[1]: Reload requested from client PID 29888 ('systemctl') (unit session-5.scope)... Dec 14 11:33:31 managed-node3 systemd[1]: Reloading... Dec 14 11:33:32 managed-node3 systemd[1]: Reloading finished in 187 ms. Dec 14 11:33:32 managed-node3 python3.12[30071]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Dec 14 11:33:32 managed-node3 systemd[1]: Starting quadlet-demo-mysql-volume.service... ░░ Subject: A start job for unit quadlet-demo-mysql-volume.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql-volume.service has begun execution. ░░ ░░ The job identifier is 1884. Dec 14 11:33:32 managed-node3 quadlet-demo-mysql-volume[30075]: time="2024-12-14T11:33:32-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" Dec 14 11:33:32 managed-node3 podman[30075]: 2024-12-14 11:33:32.664121916 -0500 EST m=+0.030237627 volume create systemd-quadlet-demo-mysql Dec 14 11:33:32 managed-node3 quadlet-demo-mysql-volume[30075]: systemd-quadlet-demo-mysql Dec 14 11:33:32 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:32 managed-node3 systemd[1]: Finished quadlet-demo-mysql-volume.service. ░░ Subject: A start job for unit quadlet-demo-mysql-volume.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql-volume.service has finished successfully. ░░ ░░ The job identifier is 1884. Dec 14 11:33:33 managed-node3 python3.12[30214]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:35 managed-node3 python3.12[30347]: ansible-file Invoked with path=/tmp/quadlet_demo state=directory owner=root group=root mode=0777 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:35 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:38 managed-node3 systemd[1]: var-lib-containers-storage-overlay-compat424278761-lower\x2dmapped.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay-compat424278761-lower\x2dmapped.mount has successfully entered the 'dead' state. Dec 14 11:33:42 managed-node3 podman[30487]: 2024-12-14 11:33:42.478789036 -0500 EST m=+6.634414537 image pull dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6 Dec 14 11:33:42 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:42 managed-node3 rsyslogd[654]: imjournal: journal files changed, reloading... [v8.2408.0-2.el10 try https://www.rsyslog.com/e/0 ] Dec 14 11:33:42 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:42 managed-node3 python3.12[30806]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:43 managed-node3 python3.12[30937]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:33:43 managed-node3 python3.12[31042]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/systemd/quadlet-demo-mysql.container owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1734194023.0743153-14590-125136684732491/.source.container _original_basename=.be5euwc5 follow=False checksum=ca62b2ad3cc9afb5b5371ebbf797b9bc4fd7edd4 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:44 managed-node3 python3.12[31173]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:33:44 managed-node3 systemd[1]: Reload requested from client PID 31174 ('systemctl') (unit session-5.scope)... Dec 14 11:33:44 managed-node3 systemd[1]: Reloading... Dec 14 11:33:44 managed-node3 systemd[1]: Reloading finished in 189 ms. Dec 14 11:33:44 managed-node3 python3.12[31357]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Dec 14 11:33:44 managed-node3 systemd[1]: Starting quadlet-demo-mysql.service... ░░ Subject: A start job for unit quadlet-demo-mysql.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql.service has begun execution. ░░ ░░ The job identifier is 1968. Dec 14 11:33:45 managed-node3 quadlet-demo-mysql[31361]: time="2024-12-14T11:33:45-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" Dec 14 11:33:45 managed-node3 podman[31361]: 2024-12-14 11:33:45.059579002 -0500 EST m=+0.044945043 container create e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Dec 14 11:33:45 managed-node3 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1085] manager: (podman1): new Bridge device (/org/freedesktop/NetworkManager/Devices/3) Dec 14 11:33:45 managed-node3 kernel: podman1: port 1(veth0) entered blocking state Dec 14 11:33:45 managed-node3 kernel: podman1: port 1(veth0) entered disabled state Dec 14 11:33:45 managed-node3 kernel: veth0: entered allmulticast mode Dec 14 11:33:45 managed-node3 kernel: veth0: entered promiscuous mode Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1254] manager: (veth0): new Veth device (/org/freedesktop/NetworkManager/Devices/4) Dec 14 11:33:45 managed-node3 kernel: podman1: port 1(veth0) entered blocking state Dec 14 11:33:45 managed-node3 kernel: podman1: port 1(veth0) entered forwarding state Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1290] device (veth0): carrier: link connected Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1292] device (podman1): carrier: link connected Dec 14 11:33:45 managed-node3 podman[31361]: 2024-12-14 11:33:45.043155252 -0500 EST m=+0.028521433 image pull dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6 Dec 14 11:33:45 managed-node3 (udev-worker)[31378]: Network interface NamePolicy= disabled on kernel command line. Dec 14 11:33:45 managed-node3 (udev-worker)[31380]: Network interface NamePolicy= disabled on kernel command line. Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1732] device (podman1): state change: unmanaged -> unavailable (reason 'connection-assumed', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1741] device (podman1): state change: unavailable -> disconnected (reason 'connection-assumed', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1748] device (podman1): Activation: starting connection 'podman1' (b4057425-e67a-485d-8a7a-63f5d24af0ff) Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1751] device (podman1): state change: disconnected -> prepare (reason 'none', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1754] device (podman1): state change: prepare -> config (reason 'none', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1756] device (podman1): state change: config -> ip-config (reason 'none', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1759] device (podman1): state change: ip-config -> ip-check (reason 'none', managed-type: 'external') Dec 14 11:33:45 managed-node3 systemd[1]: Starting NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 2055. Dec 14 11:33:45 managed-node3 systemd[1]: Started NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 2055. Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.2268] device (podman1): state change: ip-check -> secondaries (reason 'none', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.2272] device (podman1): state change: secondaries -> activated (reason 'none', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.2282] device (podman1): Activation: successful, device activated. Dec 14 11:33:45 managed-node3 systemd[1]: Started run-r7fbb015e72674f0ba85c639c3ed71a6f.scope - /usr/libexec/podman/aardvark-dns --config /run/containers/networks/aardvark-dns -p 53 run. ░░ Subject: A start job for unit run-r7fbb015e72674f0ba85c639c3ed71a6f.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r7fbb015e72674f0ba85c639c3ed71a6f.scope has finished successfully. ░░ ░░ The job identifier is 2134. Dec 14 11:33:45 managed-node3 systemd[1]: Started e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72-7b63df59ce7fdf47.timer - /usr/bin/podman healthcheck run e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72. ░░ Subject: A start job for unit e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72-7b63df59ce7fdf47.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72-7b63df59ce7fdf47.timer has finished successfully. ░░ ░░ The job identifier is 2140. Dec 14 11:33:45 managed-node3 podman[31361]: 2024-12-14 11:33:45.31271762 -0500 EST m=+0.298083770 container init e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Dec 14 11:33:45 managed-node3 systemd[1]: Started quadlet-demo-mysql.service. ░░ Subject: A start job for unit quadlet-demo-mysql.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql.service has finished successfully. ░░ ░░ The job identifier is 1968. Dec 14 11:33:45 managed-node3 podman[31361]: 2024-12-14 11:33:45.338985245 -0500 EST m=+0.324351383 container start e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Dec 14 11:33:45 managed-node3 quadlet-demo-mysql[31361]: e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 Dec 14 11:33:45 managed-node3 podman[31429]: 2024-12-14 11:33:45.496049336 -0500 EST m=+0.120347512 container health_status e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, health_status=healthy, health_failing_streak=0, health_log=, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Dec 14 11:33:46 managed-node3 python3.12[31625]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:47 managed-node3 python3.12[31769]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:48 managed-node3 python3.12[31900]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:33:48 managed-node3 python3.12[32005]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734194027.8082154-14774-224715775555281/.source.yml dest=/etc/containers/systemd/envoy-proxy-configmap.yml owner=root group=0 mode=0644 _original_basename=envoy-proxy-configmap.yml follow=False checksum=d681c7d56f912150d041873e880818b22a90c188 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:49 managed-node3 python3.12[32160]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:33:49 managed-node3 systemd[1]: Reload requested from client PID 32161 ('systemctl') (unit session-5.scope)... Dec 14 11:33:49 managed-node3 systemd[1]: Reloading... Dec 14 11:33:49 managed-node3 systemd[1]: Reloading finished in 206 ms. Dec 14 11:33:49 managed-node3 systemd[1]: Starting logrotate.service - Rotate log files... ░░ Subject: A start job for unit logrotate.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.service has begun execution. ░░ ░░ The job identifier is 2296. Dec 14 11:33:49 managed-node3 systemd[1]: logrotate.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit logrotate.service has successfully entered the 'dead' state. Dec 14 11:33:49 managed-node3 systemd[1]: Finished logrotate.service - Rotate log files. ░░ Subject: A start job for unit logrotate.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.service has finished successfully. ░░ ░░ The job identifier is 2296. Dec 14 11:33:50 managed-node3 python3.12[32349]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:52 managed-node3 python3.12[32519]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:52 managed-node3 python3.12[32650]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo.yml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:33:53 managed-node3 python3.12[32755]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/systemd/quadlet-demo.yml owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1734194032.2620435-14941-256399581694956/.source.yml _original_basename=.mz1rbt_9 follow=False checksum=998dccde0483b1654327a46ddd89cbaa47650370 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:53 managed-node3 python3.12[32893]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:33:53 managed-node3 systemd[1]: Reload requested from client PID 32894 ('systemctl') (unit session-5.scope)... Dec 14 11:33:53 managed-node3 systemd[1]: Reloading... Dec 14 11:33:53 managed-node3 systemd[1]: Reloading finished in 214 ms. Dec 14 11:33:54 managed-node3 python3.12[33078]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:55 managed-node3 systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Dec 14 11:33:55 managed-node3 python3.12[33235]: ansible-slurp Invoked with path=/etc/containers/systemd/quadlet-demo.yml src=/etc/containers/systemd/quadlet-demo.yml Dec 14 11:33:56 managed-node3 python3.12[33366]: ansible-file Invoked with path=/tmp/httpd3 state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:57 managed-node3 python3.12[33497]: ansible-file Invoked with path=/tmp/httpd3-create state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:34:09 managed-node3 podman[33636]: 2024-12-14 11:34:09.806030447 -0500 EST m=+12.237925783 image pull fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b quay.io/linux-system-roles/wordpress:4.8-apache Dec 14 11:34:15 managed-node3 podman[34066]: 2024-12-14 11:34:15.170385157 -0500 EST m=+4.909223490 image pull 5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d quay.io/linux-system-roles/envoyproxy:v1.25.0 Dec 14 11:34:15 managed-node3 python3.12[34337]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:34:15 managed-node3 podman[34446]: 2024-12-14 11:34:15.95217748 -0500 EST m=+0.115361640 container health_status e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, health_status=healthy, health_failing_streak=0, health_log=, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Dec 14 11:34:16 managed-node3 python3.12[34474]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo.kube follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:34:16 managed-node3 python3.12[34589]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734194055.7196329-15462-153591703695286/.source.kube dest=/etc/containers/systemd/quadlet-demo.kube owner=root group=0 mode=0644 _original_basename=quadlet-demo.kube follow=False checksum=7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:34:16 managed-node3 python3.12[34720]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:34:16 managed-node3 systemd[1]: Reload requested from client PID 34721 ('systemctl') (unit session-5.scope)... Dec 14 11:34:16 managed-node3 systemd[1]: Reloading... Dec 14 11:34:17 managed-node3 systemd[1]: Reloading finished in 210 ms. Dec 14 11:34:17 managed-node3 python3.12[34904]: ansible-systemd Invoked with name=quadlet-demo.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Dec 14 11:34:17 managed-node3 systemd[1]: Starting quadlet-demo.service... ░░ Subject: A start job for unit quadlet-demo.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo.service has begun execution. ░░ ░░ The job identifier is 2452. Dec 14 11:34:17 managed-node3 quadlet-demo[34908]: time="2024-12-14T11:34:17-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" Dec 14 11:34:17 managed-node3 quadlet-demo[34908]: Pods stopped: Dec 14 11:34:17 managed-node3 quadlet-demo[34908]: Pods removed: Dec 14 11:34:17 managed-node3 quadlet-demo[34908]: Secrets removed: Dec 14 11:34:17 managed-node3 quadlet-demo[34908]: Volumes removed: Dec 14 11:34:17 managed-node3 podman[34908]: 2024-12-14 11:34:17.734699606 -0500 EST m=+0.029844560 volume create wp-pv-claim Dec 14 11:34:17 managed-node3 quadlet-demo[34908]: Error: image for service container: building local pause image: finding pause binary: exec: "catatonit": executable file not found in $PATH Dec 14 11:34:17 managed-node3 systemd[1]: quadlet-demo.service: Main process exited, code=exited, status=125/n/a ░░ Subject: Unit process exited ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ An ExecStart= process belonging to unit quadlet-demo.service has exited. ░░ ░░ The process' exit code is 'exited' and its exit status is 125. Dec 14 11:34:17 managed-node3 quadlet-demo[34916]: time="2024-12-14T11:34:17-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" Dec 14 11:34:17 managed-node3 quadlet-demo[34916]: Pods stopped: Dec 14 11:34:17 managed-node3 quadlet-demo[34916]: Pods removed: Dec 14 11:34:17 managed-node3 quadlet-demo[34916]: Secrets removed: Dec 14 11:34:17 managed-node3 quadlet-demo[34916]: Volumes removed: Dec 14 11:34:17 managed-node3 systemd[1]: quadlet-demo.service: Failed with result 'exit-code'. ░░ Subject: Unit failed ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit quadlet-demo.service has entered the 'failed' state with result 'exit-code'. Dec 14 11:34:17 managed-node3 systemd[1]: Failed to start quadlet-demo.service. ░░ Subject: A start job for unit quadlet-demo.service has failed ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo.service has finished with a failure. ░░ ░░ The job identifier is 2452 and the job result is failed. Dec 14 11:34:18 managed-node3 python3.12[35056]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None TASK [Check] ******************************************************************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:148 Saturday 14 December 2024 11:34:18 -0500 (0:00:00.458) 0:01:27.351 ***** ok: [managed-node3] => { "changed": false, "cmd": [ "podman", "ps", "-a" ], "delta": "0:00:00.038665", "end": "2024-12-14 11:34:18.649886", "rc": 0, "start": "2024-12-14 11:34:18.611221" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES e8a2f6c7683e quay.io/linux-system-roles/mysql:5.6 mysqld 33 seconds ago Up 33 seconds (healthy) 3306/tcp quadlet-demo-mysql STDERR: time="2024-12-14T11:34:18-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" TASK [Check pods] ************************************************************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:152 Saturday 14 December 2024 11:34:18 -0500 (0:00:00.480) 0:01:27.832 ***** ok: [managed-node3] => { "changed": false, "cmd": [ "podman", "pod", "ps", "--ctr-ids", "--ctr-names", "--ctr-status" ], "delta": "0:00:00.028166", "end": "2024-12-14 11:34:19.142158", "failed_when_result": false, "rc": 0, "start": "2024-12-14 11:34:19.113992" } STDOUT: POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS STDERR: time="2024-12-14T11:34:19-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" TASK [Check systemd] *********************************************************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:157 Saturday 14 December 2024 11:34:19 -0500 (0:00:00.416) 0:01:28.248 ***** ok: [managed-node3] => { "changed": false, "cmd": "set -euo pipefail; systemctl list-units --all | grep quadlet", "delta": "0:00:00.013899", "end": "2024-12-14 11:34:19.527608", "failed_when_result": false, "rc": 0, "start": "2024-12-14 11:34:19.513709" } STDOUT: quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service quadlet-demo-mysql.service loaded active running quadlet-demo-mysql.service quadlet-demo-network.service loaded active exited quadlet-demo-network.service ● quadlet-demo.service loaded failed failed quadlet-demo.service TASK [LS] ********************************************************************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:165 Saturday 14 December 2024 11:34:19 -0500 (0:00:00.384) 0:01:28.633 ***** ok: [managed-node3] => { "changed": false, "cmd": [ "ls", "-alrtF", "/etc/systemd/system" ], "delta": "0:00:00.004563", "end": "2024-12-14 11:34:19.902569", "failed_when_result": false, "rc": 0, "start": "2024-12-14 11:34:19.898006" } STDOUT: total 12 drwxr-xr-x. 5 root root 47 Dec 3 00:45 ../ drwxr-xr-x. 2 root root 32 Dec 3 00:45 getty.target.wants/ lrwxrwxrwx. 1 root root 43 Dec 3 00:45 dbus.service -> /usr/lib/systemd/system/dbus-broker.service lrwxrwxrwx. 1 root root 37 Dec 3 00:45 ctrl-alt-del.target -> /usr/lib/systemd/system/reboot.target lrwxrwxrwx. 1 root root 57 Dec 3 00:45 dbus-org.freedesktop.nm-dispatcher.service -> /usr/lib/systemd/system/NetworkManager-dispatcher.service drwxr-xr-x. 2 root root 48 Dec 3 00:45 network-online.target.wants/ drwxr-xr-x. 2 root root 76 Dec 3 00:45 timers.target.wants/ drwxr-xr-x. 2 root root 38 Dec 3 00:46 dev-virtio\x2dports-org.qemu.guest_agent.0.device.wants/ lrwxrwxrwx. 1 root root 41 Dec 3 00:48 default.target -> /usr/lib/systemd/system/multi-user.target drwxr-xr-x. 2 root root 31 Dec 3 02:57 remote-fs.target.wants/ drwxr-xr-x. 2 root root 119 Dec 3 02:58 cloud-init.target.wants/ drwxr-xr-x. 2 root root 113 Dec 3 02:59 sockets.target.wants/ drwxr-xr-x. 2 root root 4096 Dec 3 02:59 sysinit.target.wants/ drwxr-xr-x. 2 root root 4096 Dec 14 11:33 multi-user.target.wants/ lrwxrwxrwx. 1 root root 41 Dec 14 11:33 dbus-org.fedoraproject.FirewallD1.service -> /usr/lib/systemd/system/firewalld.service drwxr-xr-x. 11 root root 4096 Dec 14 11:33 ./ TASK [Cleanup] ***************************************************************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:172 Saturday 14 December 2024 11:34:19 -0500 (0:00:00.391) 0:01:29.025 ***** included: fedora.linux_system_roles.podman for managed-node3 TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Saturday 14 December 2024 11:34:20 -0500 (0:00:00.120) 0:01:29.145 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Saturday 14 December 2024 11:34:20 -0500 (0:00:00.085) 0:01:29.230 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Saturday 14 December 2024 11:34:20 -0500 (0:00:00.060) 0:01:29.291 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Saturday 14 December 2024 11:34:20 -0500 (0:00:00.047) 0:01:29.338 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Saturday 14 December 2024 11:34:20 -0500 (0:00:00.034) 0:01:29.373 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Saturday 14 December 2024 11:34:20 -0500 (0:00:00.037) 0:01:29.410 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Saturday 14 December 2024 11:34:20 -0500 (0:00:00.037) 0:01:29.448 ***** ok: [managed-node3] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed-node3] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node3] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } ok: [managed-node3] => (item=CentOS_10.yml) => { "ansible_facts": { "__podman_packages": [ "iptables-nft", "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_10.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_10.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Saturday 14 December 2024 11:34:20 -0500 (0:00:00.073) 0:01:29.522 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Saturday 14 December 2024 11:34:21 -0500 (0:00:00.780) 0:01:30.302 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Saturday 14 December 2024 11:34:21 -0500 (0:00:00.038) 0:01:30.340 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages))", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Saturday 14 December 2024 11:34:21 -0500 (0:00:00.042) 0:01:30.382 ***** skipping: [managed-node3] => { "false_condition": "__podman_is_transactional | d(false)" } TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Saturday 14 December 2024 11:34:21 -0500 (0:00:00.030) 0:01:30.413 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Saturday 14 December 2024 11:34:21 -0500 (0:00:00.072) 0:01:30.485 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Saturday 14 December 2024 11:34:21 -0500 (0:00:00.031) 0:01:30.517 ***** ok: [managed-node3] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.025553", "end": "2024-12-14 11:34:21.814490", "rc": 0, "start": "2024-12-14 11:34:21.788937" } STDOUT: podman version 5.3.1 STDERR: time="2024-12-14T11:34:21-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Saturday 14 December 2024 11:34:21 -0500 (0:00:00.402) 0:01:30.920 ***** ok: [managed-node3] => { "ansible_facts": { "podman_version": "5.3.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Saturday 14 December 2024 11:34:21 -0500 (0:00:00.032) 0:01:30.953 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Saturday 14 December 2024 11:34:21 -0500 (0:00:00.029) 0:01:30.983 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Saturday 14 December 2024 11:34:21 -0500 (0:00:00.035) 0:01:31.018 ***** META: end_host conditional evaluated to False, continuing execution for managed-node3 skipping: [managed-node3] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node3" } MSG: end_host conditional evaluated to false, continuing execution for managed-node3 TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Saturday 14 December 2024 11:34:22 -0500 (0:00:00.037) 0:01:31.056 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__has_type_pod or __has_pod_file_ext or __has_pod_file_src_ext or __has_pod_template_src_ext or __has_pod_template_src_ext_j2", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:96 Saturday 14 December 2024 11:34:22 -0500 (0:00:00.059) 0:01:31.115 ***** META: end_host conditional evaluated to False, continuing execution for managed-node3 skipping: [managed-node3] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node3" } MSG: end_host conditional evaluated to false, continuing execution for managed-node3 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:109 Saturday 14 December 2024 11:34:22 -0500 (0:00:00.074) 0:01:31.190 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 14 December 2024 11:34:22 -0500 (0:00:00.069) 0:01:31.259 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 14 December 2024 11:34:22 -0500 (0:00:00.041) 0:01:31.301 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 14 December 2024 11:34:22 -0500 (0:00:00.037) 0:01:31.338 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 14 December 2024 11:34:22 -0500 (0:00:00.042) 0:01:31.381 ***** ok: [managed-node3] => { "changed": false, "stat": { "atime": 1734193882.6705027, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1734193862.9755826, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 8859182, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "2878164177", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 14 December 2024 11:34:22 -0500 (0:00:00.390) 0:01:31.772 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:44 Saturday 14 December 2024 11:34:22 -0500 (0:00:00.054) 0:01:31.826 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Saturday 14 December 2024 11:34:22 -0500 (0:00:00.050) 0:01:31.876 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:62 Saturday 14 December 2024 11:34:22 -0500 (0:00:00.051) 0:01:31.928 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:67 Saturday 14 December 2024 11:34:22 -0500 (0:00:00.048) 0:01:31.976 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:72 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.109) 0:01:32.086 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:82 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.050) 0:01:32.137 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:89 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.052) 0:01:32.189 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:115 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.048) 0:01:32.238 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:124 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.066) 0:01:32.305 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.098) 0:01:32.404 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.046) 0:01:32.451 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:127 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.055) 0:01:32.506 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.098) 0:01:32.606 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.046) 0:01:32.652 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:130 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.055) 0:01:32.707 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.103) 0:01:32.811 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.051) 0:01:32.863 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:133 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.051) 0:01:32.914 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Saturday 14 December 2024 11:34:23 -0500 (0:00:00.103) 0:01:33.017 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Saturday 14 December 2024 11:34:24 -0500 (0:00:00.050) 0:01:33.068 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Saturday 14 December 2024 11:34:24 -0500 (0:00:00.048) 0:01:33.116 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Saturday 14 December 2024 11:34:24 -0500 (0:00:00.049) 0:01:33.166 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:139 Saturday 14 December 2024 11:34:24 -0500 (0:00:00.111) 0:01:33.278 ***** included: fedora.linux_system_roles.firewall for managed-node3 TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Saturday 14 December 2024 11:34:24 -0500 (0:00:00.162) 0:01:33.440 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed-node3 TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Saturday 14 December 2024 11:34:24 -0500 (0:00:00.090) 0:01:33.531 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Saturday 14 December 2024 11:34:24 -0500 (0:00:00.057) 0:01:33.588 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Saturday 14 December 2024 11:34:24 -0500 (0:00:00.052) 0:01:33.640 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __firewall_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22 Saturday 14 December 2024 11:34:24 -0500 (0:00:00.050) 0:01:33.690 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __firewall_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27 Saturday 14 December 2024 11:34:24 -0500 (0:00:00.052) 0:01:33.743 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __firewall_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 Saturday 14 December 2024 11:34:24 -0500 (0:00:00.050) 0:01:33.794 ***** ok: [managed-node3] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: firewalld TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43 Saturday 14 December 2024 11:34:25 -0500 (0:00:00.740) 0:01:34.534 ***** skipping: [managed-node3] => { "false_condition": "__firewall_is_transactional | d(false)" } TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48 Saturday 14 December 2024 11:34:25 -0500 (0:00:00.031) 0:01:34.566 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53 Saturday 14 December 2024 11:34:25 -0500 (0:00:00.030) 0:01:34.596 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Saturday 14 December 2024 11:34:25 -0500 (0:00:00.035) 0:01:34.632 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Saturday 14 December 2024 11:34:25 -0500 (0:00:00.048) 0:01:34.680 ***** skipping: [managed-node3] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall_disable_conflicting_services | bool", "item": "ufw", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Saturday 14 December 2024 11:34:25 -0500 (0:00:00.060) 0:01:34.741 ***** ok: [managed-node3] => { "changed": false, "name": "firewalld", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sat 2024-12-14 11:33:17 EST", "ActiveEnterTimestampMonotonic": "418447836", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "basic.target dbus-broker.service dbus.socket sysinit.target polkit.service system.slice", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sat 2024-12-14 11:33:16 EST", "AssertTimestampMonotonic": "418197394", "Before": "shutdown.target network-pre.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "425588000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2024-12-14 11:33:16 EST", "ConditionTimestampMonotonic": "418197391", "ConfigurationDirectoryMode": "0755", "Conflicts": "ip6tables.service ipset.service iptables.service ebtables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4759", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryHigh": "3698237440", "EffectiveMemoryMax": "3698237440", "EffectiveMemoryNodes": "0", "EffectiveTasksMax": "22361", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Sat 2024-12-14 11:33:17 EST", "ExecMainHandoffTimestampMonotonic": "418226069", "ExecMainPID": "27637", "ExecMainStartTimestamp": "Sat 2024-12-14 11:33:16 EST", "ExecMainStartTimestampMonotonic": "418200543", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2024-12-14 11:33:17 EST", "InactiveExitTimestampMonotonic": "418201207", "InvocationID": "4e8dd4b4cbce4e94a93f2e02e428ae35", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "27637", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "2542981120", "MemoryCurrent": "32481280", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "33775616", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "0", "MemorySwapMax": "infinity", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target dbus.socket system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestamp": "Sat 2024-12-14 11:34:17 EST", "StateChangeTimestampMonotonic": "478348193", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22361", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Saturday 14 December 2024 11:34:26 -0500 (0:00:00.570) 0:01:35.311 ***** ok: [managed-node3] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "AccessSELinuxContext": "system_u:object_r:firewalld_unit_file_t:s0", "ActiveEnterTimestamp": "Sat 2024-12-14 11:33:17 EST", "ActiveEnterTimestampMonotonic": "418447836", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "basic.target dbus-broker.service dbus.socket sysinit.target polkit.service system.slice", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sat 2024-12-14 11:33:16 EST", "AssertTimestampMonotonic": "418197394", "Before": "shutdown.target network-pre.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "425588000", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_tty_config cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2024-12-14 11:33:16 EST", "ConditionTimestampMonotonic": "418197391", "ConfigurationDirectoryMode": "0755", "Conflicts": "ip6tables.service ipset.service iptables.service ebtables.service shutdown.target", "ControlGroup": "/system.slice/firewalld.service", "ControlGroupId": "4759", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DeviceAllow": "char-rtc r", "DevicePolicy": "closed", "Documentation": "\"man:firewalld(1)\"", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryHigh": "3698237440", "EffectiveMemoryMax": "3698237440", "EffectiveMemoryNodes": "0", "EffectiveTasksMax": "22361", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Sat 2024-12-14 11:33:17 EST", "ExecMainHandoffTimestampMonotonic": "418226069", "ExecMainPID": "27637", "ExecMainStartTimestamp": "Sat 2024-12-14 11:33:16 EST", "ExecMainStartTimestampMonotonic": "418200543", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Sat 2024-12-14 11:33:17 EST", "InactiveExitTimestampMonotonic": "418201207", "InvocationID": "4e8dd4b4cbce4e94a93f2e02e428ae35", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "yes", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "27637", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "2558504960", "MemoryCurrent": "32481280", "MemoryDenyWriteExecute": "yes", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "33775616", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "0", "MemorySwapMax": "infinity", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "firewalld.service dbus-org.fedoraproject.FirewallD1.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "yes", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "yes", "ProtectControlGroups": "yes", "ProtectHome": "yes", "ProtectHostname": "yes", "ProtectKernelLogs": "yes", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "yes", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target dbus.socket system.slice", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "yes", "RestrictSUIDSGID": "yes", "Result": "success", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestamp": "Sat 2024-12-14 11:34:17 EST", "StateChangeTimestampMonotonic": "478348193", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallArchitectures": "native", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22361", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Saturday 14 December 2024 11:34:26 -0500 (0:00:00.533) 0:01:35.845 ***** ok: [managed-node3] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/bin/python3.12", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Saturday 14 December 2024 11:34:26 -0500 (0:00:00.042) 0:01:35.888 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Saturday 14 December 2024 11:34:26 -0500 (0:00:00.029) 0:01:35.917 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Saturday 14 December 2024 11:34:26 -0500 (0:00:00.082) 0:01:36.000 ***** ok: [managed-node3] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" } } ok: [managed-node3] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Saturday 14 December 2024 11:34:27 -0500 (0:00:01.012) 0:01:37.012 ***** skipping: [managed-node3] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "firewall | length == 1", "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed-node3] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.056) 0:01:37.069 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "firewall | length == 1", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.043) 0:01:37.113 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.038) 0:01:37.151 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "firewall == None or firewall | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.032) 0:01:37.184 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.032) 0:01:37.216 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__firewall_previous_replaced | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.030) 0:01:37.247 ***** skipping: [managed-node3] => { "false_condition": "__firewall_previous_replaced | bool" } TASK [Manage selinux for specified ports] ************************************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:146 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.046) 0:01:37.293 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:153 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.037) 0:01:37.330 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:157 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.038) 0:01:37.368 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:166 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.047) 0:01:37.416 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:175 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.031) 0:01:37.448 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed-node3 => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.107) 0:01:37.555 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.035) 0:01:37.591 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.038) 0:01:37.629 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.049) 0:01:37.678 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.029) 0:01:37.708 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.035) 0:01:37.743 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.106) 0:01:37.850 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Saturday 14 December 2024 11:34:28 -0500 (0:00:00.050) 0:01:37.900 ***** changed: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Saturday 14 December 2024 11:34:29 -0500 (0:00:00.461) 0:01:38.362 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Saturday 14 December 2024 11:34:29 -0500 (0:00:00.055) 0:01:38.418 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Saturday 14 December 2024 11:34:29 -0500 (0:00:00.063) 0:01:38.481 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:34:29 -0500 (0:00:00.081) 0:01:38.563 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:34:29 -0500 (0:00:00.039) 0:01:38.603 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:34:29 -0500 (0:00:00.033) 0:01:38.636 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Saturday 14 December 2024 11:34:29 -0500 (0:00:00.037) 0:01:38.673 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Saturday 14 December 2024 11:34:29 -0500 (0:00:00.034) 0:01:38.707 ***** changed: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Saturday 14 December 2024 11:34:30 -0500 (0:00:00.422) 0:01:39.130 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Saturday 14 December 2024 11:34:30 -0500 (0:00:00.042) 0:01:39.172 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Saturday 14 December 2024 11:34:30 -0500 (0:00:00.041) 0:01:39.213 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:34:30 -0500 (0:00:00.049) 0:01:39.262 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:34:30 -0500 (0:00:00.030) 0:01:39.293 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:34:30 -0500 (0:00:00.031) 0:01:39.325 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Saturday 14 December 2024 11:34:30 -0500 (0:00:00.030) 0:01:39.355 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Saturday 14 December 2024 11:34:30 -0500 (0:00:00.032) 0:01:39.387 ***** changed: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:182 Saturday 14 December 2024 11:34:30 -0500 (0:00:00.412) 0:01:39.800 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:189 Saturday 14 December 2024 11:34:30 -0500 (0:00:00.079) 0:01:39.880 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node3 => (item=(censored due to no_log)) included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node3 => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 14 December 2024 11:34:30 -0500 (0:00:00.150) 0:01:40.030 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.044) 0:01:40.075 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.040) 0:01:40.115 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_quadlet_file_src", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.034) 0:01:40.149 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.047) 0:01:40.197 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.060) 0:01:40.258 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.036) 0:01:40.294 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.034) 0:01:40.328 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.045) 0:01:40.374 ***** ok: [managed-node3] => { "changed": false, "stat": { "atime": 1734193882.6705027, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1734193862.9755826, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 8859182, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "2878164177", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.373) 0:01:40.747 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:44 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.031) 0:01:40.779 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.030) 0:01:40.809 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:62 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.032) 0:01:40.842 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:67 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.030) 0:01:40.872 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:72 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.039) 0:01:40.911 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:82 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.030) 0:01:40.942 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:89 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.031) 0:01:40.973 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 14 December 2024 11:34:31 -0500 (0:00:00.030) 0:01:41.004 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": [ "quadlet-demo.yml" ], "__podman_service_name": "quadlet-demo.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 14 December 2024 11:34:32 -0500 (0:00:00.164) 0:01:41.168 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 14 December 2024 11:34:32 -0500 (0:00:00.033) 0:01:41.202 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 14 December 2024 11:34:32 -0500 (0:00:00.030) 0:01:41.233 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:105 Saturday 14 December 2024 11:34:32 -0500 (0:00:00.074) 0:01:41.308 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:112 Saturday 14 December 2024 11:34:32 -0500 (0:00:00.040) 0:01:41.348 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Saturday 14 December 2024 11:34:32 -0500 (0:00:00.080) 0:01:41.428 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Saturday 14 December 2024 11:34:32 -0500 (0:00:00.043) 0:01:41.472 ***** changed: [managed-node3] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo.service", "state": "stopped", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "failed", "After": "quadlet-demo-network.service basic.target sysinit.target quadlet-demo-mysql.service system.slice network-online.target -.mount systemd-journald.socket", "AllowIsolate": "no", "AssertResult": "yes", "AssertTimestamp": "Sat 2024-12-14 11:34:17 EST", "AssertTimestampMonotonic": "478884455", "Before": "multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Sat 2024-12-14 11:34:17 EST", "ConditionTimestampMonotonic": "478884451", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "CoredumpReceive": "no", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "DefaultStartupMemoryLow": "0", "Delegate": "no", "Description": "quadlet-demo.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveMemoryHigh": "3698237440", "EffectiveMemoryMax": "3698237440", "EffectiveTasksMax": "22361", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo.service", "ExecMainCode": "1", "ExecMainExitTimestamp": "Sat 2024-12-14 11:34:17 EST", "ExecMainExitTimestampMonotonic": "478939421", "ExecMainHandoffTimestamp": "Sat 2024-12-14 11:34:17 EST", "ExecMainHandoffTimestampMonotonic": "478893080", "ExecMainPID": "34908", "ExecMainStartTimestamp": "Sat 2024-12-14 11:34:17 EST", "ExecMainStartTimestampMonotonic": "478885866", "ExecMainStatus": "125", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[Sat 2024-12-14 11:34:17 EST] ; stop_time=[Sat 2024-12-14 11:34:17 EST] ; pid=34908 ; code=exited ; status=125 }", "ExecStartEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; flags= ; start_time=[Sat 2024-12-14 11:34:17 EST] ; stop_time=[Sat 2024-12-14 11:34:17 EST] ; pid=34908 ; code=exited ; status=125 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[Sat 2024-12-14 11:34:17 EST] ; stop_time=[Sat 2024-12-14 11:34:17 EST] ; pid=34916 ; code=exited ; status=0 }", "ExecStopPostEx": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; flags= ; start_time=[Sat 2024-12-14 11:34:17 EST] ; stop_time=[Sat 2024-12-14 11:34:17 EST] ; pid=34916 ; code=exited ; status=0 }", "ExitType": "main", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FileDescriptorStorePreserve": "restart", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/quadlet-demo.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "quadlet-demo.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestamp": "Sat 2024-12-14 11:34:17 EST", "InactiveEnterTimestampMonotonic": "478978183", "InactiveExitTimestamp": "Sat 2024-12-14 11:34:17 EST", "InactiveExitTimestampMonotonic": "478886370", "InvocationID": "db5b269384d0437bae76c9e51d6c2b72", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "2576830464", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryKSM": "no", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemoryPeak": "[not set]", "MemoryPressureThresholdUSec": "200ms", "MemoryPressureWatch": "auto", "MemorySwapCurrent": "[not set]", "MemorySwapMax": "infinity", "MemorySwapPeak": "[not set]", "MemoryZSwapCurrent": "[not set]", "MemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MountAPIVFS": "no", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "quadlet-demo.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMPolicy": "stop", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "quadlet-demo-network.service sysinit.target -.mount quadlet-demo-mysql.service system.slice", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartKillSignal": "15", "RestartMaxDelayUSec": "infinity", "RestartMode": "normal", "RestartSteps": "0", "RestartUSec": "100ms", "RestartUSecNext": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "exit-code", "RootDirectoryStartOnly": "no", "RootEphemeral": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "SetLoginEnvironment": "no", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo.kube", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StartupMemoryHigh": "infinity", "StartupMemoryLow": "0", "StartupMemoryMax": "infinity", "StartupMemorySwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "StateChangeTimestamp": "Sat 2024-12-14 11:34:17 EST", "StateChangeTimestampMonotonic": "478978183", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "failed", "SuccessAction": "none", "SurviveFinalKillSignal": "no", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22361", "TimeoutAbortUSec": "1min 30s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "1min 30s", "TimeoutStopFailureMode": "terminate", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-online.target", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Saturday 14 December 2024 11:34:33 -0500 (0:00:00.811) 0:01:42.283 ***** ok: [managed-node3] => { "changed": false, "stat": { "atime": 1734194056.335178, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7", "ctime": 1734194056.341178, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 419430883, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1734194056.0621774, "nlink": 1, "path": "/etc/containers/systemd/quadlet-demo.kube", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 456, "uid": 0, "version": "1711308696", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Saturday 14 December 2024 11:34:33 -0500 (0:00:00.416) 0:01:42.699 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Saturday 14 December 2024 11:34:33 -0500 (0:00:00.055) 0:01:42.755 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Saturday 14 December 2024 11:34:34 -0500 (0:00:00.365) 0:01:43.120 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Saturday 14 December 2024 11:34:34 -0500 (0:00:00.049) 0:01:43.169 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Saturday 14 December 2024 11:34:34 -0500 (0:00:00.032) 0:01:43.202 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Saturday 14 December 2024 11:34:34 -0500 (0:00:00.034) 0:01:43.236 ***** changed: [managed-node3] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.kube", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Saturday 14 December 2024 11:34:34 -0500 (0:00:00.375) 0:01:43.612 ***** ok: [managed-node3] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Saturday 14 December 2024 11:34:35 -0500 (0:00:00.744) 0:01:44.356 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:99 Saturday 14 December 2024 11:34:35 -0500 (0:00:00.034) 0:01:44.391 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Saturday 14 December 2024 11:34:35 -0500 (0:00:00.043) 0:01:44.435 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:120 Saturday 14 December 2024 11:34:35 -0500 (0:00:00.033) 0:01:44.468 ***** changed: [managed-node3] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.716718", "end": "2024-12-14 11:34:36.451771", "rc": 0, "start": "2024-12-14 11:34:35.735053" } STDOUT: fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b 5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d STDERR: time="2024-12-14T11:34:35-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:131 Saturday 14 December 2024 11:34:36 -0500 (0:00:01.157) 0:01:45.625 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Saturday 14 December 2024 11:34:36 -0500 (0:00:00.061) 0:01:45.686 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Saturday 14 December 2024 11:34:36 -0500 (0:00:00.032) 0:01:45.719 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Saturday 14 December 2024 11:34:36 -0500 (0:00:00.030) 0:01:45.750 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:141 Saturday 14 December 2024 11:34:36 -0500 (0:00:00.033) 0:01:45.783 ***** ok: [managed-node3] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.031667", "end": "2024-12-14 11:34:37.089872", "rc": 0, "start": "2024-12-14 11:34:37.058205" } STDOUT: quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB STDERR: time="2024-12-14T11:34:37-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:150 Saturday 14 December 2024 11:34:37 -0500 (0:00:00.413) 0:01:46.197 ***** ok: [managed-node3] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.029915", "end": "2024-12-14 11:34:37.494176", "rc": 0, "start": "2024-12-14 11:34:37.464261" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim STDERR: time="2024-12-14T11:34:37-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:159 Saturday 14 December 2024 11:34:37 -0500 (0:00:00.410) 0:01:46.608 ***** ok: [managed-node3] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.033155", "end": "2024-12-14 11:34:37.911910", "rc": 0, "start": "2024-12-14 11:34:37.878755" } STDOUT: e8a2f6c7683e quay.io/linux-system-roles/mysql:5.6 mysqld 52 seconds ago Up 52 seconds (healthy) 3306/tcp quadlet-demo-mysql STDERR: time="2024-12-14T11:34:37-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:168 Saturday 14 December 2024 11:34:37 -0500 (0:00:00.424) 0:01:47.032 ***** ok: [managed-node3] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.029201", "end": "2024-12-14 11:34:38.338819", "rc": 0, "start": "2024-12-14 11:34:38.309618" } STDOUT: podman systemd-quadlet-demo STDERR: time="2024-12-14T11:34:38-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:177 Saturday 14 December 2024 11:34:38 -0500 (0:00:00.428) 0:01:47.461 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - pods] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:187 Saturday 14 December 2024 11:34:38 -0500 (0:00:00.416) 0:01:47.877 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:197 Saturday 14 December 2024 11:34:39 -0500 (0:00:00.431) 0:01:48.308 ***** ok: [managed-node3] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "audit-rules.service": { "name": "audit-rules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "capsule@.service": { "name": "capsule@.service", "source": "systemd", "state": "unknown", "status": "static" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dhcpcd.service": { "name": "dhcpcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dhcpcd@.service": { "name": "dhcpcd@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72-7b63df59ce7fdf47.service": { "name": "e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72-7b63df59ce7fdf47.service", "source": "systemd", "state": "stopped", "status": "transient" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fips-crypto-policy-overlay.service": { "name": "fips-crypto-policy-overlay.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "hv_kvp_daemon.service": { "name": "hv_kvp_daemon.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "ip6tables.service": { "name": "ip6tables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ipset.service": { "name": "ipset.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iptables.service": { "name": "iptables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm-devices-import.service": { "name": "lvm-devices-import.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@efi_pstore.service": { "name": "modprobe@efi_pstore.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "netavark-dhcp-proxy.service": { "name": "netavark-dhcp-proxy.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "netavark-firewalld-reload.service": { "name": "netavark-firewalld-reload.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "running", "status": "generated" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo.service": { "name": "quadlet-demo.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "quotaon-root.service": { "name": "quotaon-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "quotaon@.service": { "name": "quotaon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-battery-check.service": { "name": "systemd-battery-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-bootctl@.service": { "name": "systemd-bootctl@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-confext.service": { "name": "systemd-confext.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-creds@.service": { "name": "systemd-creds@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-clear.service": { "name": "systemd-hibernate-clear.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hibernate-resume.service": { "name": "systemd-hibernate-resume.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald-sync@.service": { "name": "systemd-journald-sync@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-pcrextend@.service": { "name": "systemd-pcrextend@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrlock-file-system.service": { "name": "systemd-pcrlock-file-system.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-pcrlock-firmware-code.service": { "name": "systemd-pcrlock-firmware-code.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-pcrlock-firmware-config.service": { "name": "systemd-pcrlock-firmware-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-pcrlock-machine-id.service": { "name": "systemd-pcrlock-machine-id.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-pcrlock-make-policy.service": { "name": "systemd-pcrlock-make-policy.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-pcrlock-secureboot-authority.service": { "name": "systemd-pcrlock-secureboot-authority.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-pcrlock-secureboot-policy.service": { "name": "systemd-pcrlock-secureboot-policy.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-pcrlock@.service": { "name": "systemd-pcrlock@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck-root.service": { "name": "systemd-quotacheck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-quotacheck@.service": { "name": "systemd-quotacheck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-soft-reboot.service": { "name": "systemd-soft-reboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-sysext@.service": { "name": "systemd-sysext@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev-early.service": { "name": "systemd-tmpfiles-setup-dev-early.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tpm2-setup-early.service": { "name": "systemd-tpm2-setup-early.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tpm2-setup.service": { "name": "systemd-tpm2-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-load-credentials.service": { "name": "systemd-udev-load-credentials.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:116 Saturday 14 December 2024 11:34:41 -0500 (0:00:02.186) 0:01:50.495 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Saturday 14 December 2024 11:34:41 -0500 (0:00:00.034) 0:01:50.529 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n", "__podman_quadlet_template_src": "quadlet-demo.yml.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Saturday 14 December 2024 11:34:41 -0500 (0:00:00.097) 0:01:50.627 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Saturday 14 December 2024 11:34:41 -0500 (0:00:00.045) 0:01:50.672 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_quadlet_str", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Saturday 14 December 2024 11:34:41 -0500 (0:00:00.036) 0:01:50.709 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Saturday 14 December 2024 11:34:41 -0500 (0:00:00.052) 0:01:50.761 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Saturday 14 December 2024 11:34:41 -0500 (0:00:00.121) 0:01:50.882 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Saturday 14 December 2024 11:34:41 -0500 (0:00:00.047) 0:01:50.930 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Saturday 14 December 2024 11:34:41 -0500 (0:00:00.060) 0:01:50.990 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.091) 0:01:51.082 ***** ok: [managed-node3] => { "changed": false, "stat": { "atime": 1734193882.6705027, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "89ab10a2a8fa81bcc0c1df0058f200469ce46f97", "ctime": 1734193862.9755826, "dev": 51714, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 8859182, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1730678400.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15744, "uid": 0, "version": "2878164177", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.433) 0:01:51.516 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:44 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.058) 0:01:51.574 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:49 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.053) 0:01:51.628 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:62 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.039) 0:01:51.667 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:67 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.042) 0:01:51.709 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:72 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.038) 0:01:51.748 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:82 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.037) 0:01:51.786 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:89 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.034) 0:01:51.820 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.032) 0:01:51.853 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.059) 0:01:51.912 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.040) 0:01:51.952 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Saturday 14 December 2024 11:34:42 -0500 (0:00:00.041) 0:01:51.994 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:105 Saturday 14 December 2024 11:34:43 -0500 (0:00:00.177) 0:01:52.171 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:112 Saturday 14 December 2024 11:34:43 -0500 (0:00:00.046) 0:01:52.218 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Saturday 14 December 2024 11:34:43 -0500 (0:00:00.079) 0:01:52.298 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Saturday 14 December 2024 11:34:43 -0500 (0:00:00.034) 0:01:52.332 ***** skipping: [managed-node3] => { "changed": false, "false_condition": "__podman_service_name | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Saturday 14 December 2024 11:34:43 -0500 (0:00:00.036) 0:01:52.369 ***** ok: [managed-node3] => { "changed": false, "stat": { "atime": 1734194035.772143, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "998dccde0483b1654327a46ddd89cbaa47650370", "ctime": 1734194033.1691387, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 364904676, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1734194032.7701378, "nlink": 1, "path": "/etc/containers/systemd/quadlet-demo.yml", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1605, "uid": 0, "version": "3188214184", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Saturday 14 December 2024 11:34:43 -0500 (0:00:00.410) 0:01:52.779 ***** included: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed-node3 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Saturday 14 December 2024 11:34:43 -0500 (0:00:00.095) 0:01:52.875 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Saturday 14 December 2024 11:34:44 -0500 (0:00:00.420) 0:01:53.295 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Saturday 14 December 2024 11:34:44 -0500 (0:00:00.059) 0:01:53.355 ***** ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Saturday 14 December 2024 11:34:44 -0500 (0:00:00.078) 0:01:53.433 ***** ok: [managed-node3] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Saturday 14 December 2024 11:34:44 -0500 (0:00:00.055) 0:01:53.489 ***** changed: [managed-node3] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Saturday 14 December 2024 11:34:44 -0500 (0:00:00.406) 0:01:53.896 ***** ok: [managed-node3] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Saturday 14 December 2024 11:34:45 -0500 (0:00:00.755) 0:01:54.651 ***** skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:99 Saturday 14 December 2024 11:34:45 -0500 (0:00:00.036) 0:01:54.688 ***** failed: [managed-node3] (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } failed: [managed-node3] (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [managed-node3] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } fatal: [managed-node3]: FAILED! => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [Debug] ******************************************************************* task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:199 Saturday 14 December 2024 11:34:46 -0500 (0:00:01.332) 0:01:56.021 ***** ok: [managed-node3] => { "changed": false, "cmd": "exec 1>&2\nset -x\nset -o pipefail\nsystemctl list-units --plain -l --all | grep quadlet || :\nsystemctl list-unit-files --all | grep quadlet || :\nsystemctl list-units --plain --failed -l --all | grep quadlet || :\n", "delta": "0:00:00.381790", "end": "2024-12-14 11:34:47.677674", "rc": 0, "start": "2024-12-14 11:34:47.295884" } STDERR: + set -o pipefail + systemctl list-units --plain -l --all + grep quadlet quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service quadlet-demo-mysql.service loaded active running quadlet-demo-mysql.service quadlet-demo-network.service loaded active exited quadlet-demo-network.service quadlet-demo.service not-found failed failed quadlet-demo.service + systemctl list-unit-files --all + grep quadlet quadlet-demo-mysql-volume.service generated - quadlet-demo-mysql.service generated - quadlet-demo-network.service generated - + systemctl list-units --plain --failed -l --all + grep quadlet quadlet-demo.service not-found failed failed quadlet-demo.service TASK [Get journald] ************************************************************ task path: /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:209 Saturday 14 December 2024 11:34:47 -0500 (0:00:00.787) 0:01:56.809 ***** fatal: [managed-node3]: FAILED! => { "changed": false, "cmd": [ "journalctl", "-ex" ], "delta": "0:00:00.026734", "end": "2024-12-14 11:34:48.098878", "failed_when_result": true, "rc": 0, "start": "2024-12-14 11:34:48.072144" } STDOUT: Dec 14 11:26:32 localhost augenrules[617]: failure 1 Dec 14 11:26:32 localhost augenrules[617]: pid 0 Dec 14 11:26:32 localhost augenrules[617]: rate_limit 0 Dec 14 11:26:32 localhost augenrules[617]: backlog_limit 8192 Dec 14 11:26:32 localhost augenrules[617]: lost 0 Dec 14 11:26:32 localhost augenrules[617]: backlog 0 Dec 14 11:26:32 localhost augenrules[617]: backlog_wait_time 60000 Dec 14 11:26:32 localhost augenrules[617]: backlog_wait_time_actual 0 Dec 14 11:26:32 localhost systemd[1]: audit-rules.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit audit-rules.service has successfully entered the 'dead' state. Dec 14 11:26:32 localhost systemd[1]: Finished audit-rules.service - Load Audit Rules. ░░ Subject: A start job for unit audit-rules.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit audit-rules.service has finished successfully. ░░ ░░ The job identifier is 248. Dec 14 11:26:32 localhost systemd[1]: Starting auditd.service - Security Audit Logging Service... ░░ Subject: A start job for unit auditd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has begun execution. ░░ ░░ The job identifier is 247. Dec 14 11:26:33 localhost auditd[628]: No plugins found, not dispatching events Dec 14 11:26:33 localhost auditd[628]: Init complete, auditd 4.0 listening for events (startup state enable) Dec 14 11:26:33 localhost systemd[1]: Started auditd.service - Security Audit Logging Service. ░░ Subject: A start job for unit auditd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit auditd.service has finished successfully. ░░ ░░ The job identifier is 247. Dec 14 11:26:33 localhost kernel: cirrus 0000:00:02.0: vgaarb: deactivate vga console Dec 14 11:26:33 localhost kernel: Console: switching to colour dummy device 80x25 Dec 14 11:26:33 localhost kernel: [drm] Initialized cirrus 2.0.0 for 0000:00:02.0 on minor 0 Dec 14 11:26:33 localhost kernel: fbcon: cirrusdrmfb (fb0) is primary device Dec 14 11:26:33 localhost kernel: Console: switching to colour frame buffer device 128x48 Dec 14 11:26:33 localhost kernel: cirrus 0000:00:02.0: [drm] fb0: cirrusdrmfb frame buffer device Dec 14 11:26:33 localhost systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... ░░ Subject: A start job for unit systemd-update-utmp.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp.service has begun execution. ░░ ░░ The job identifier is 236. Dec 14 11:26:33 localhost systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... ░░ Subject: A start job for unit systemd-vconsole-setup.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-vconsole-setup.service has begun execution. ░░ ░░ The job identifier is 300. Dec 14 11:26:33 localhost systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. ░░ Subject: A start job for unit systemd-update-utmp.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp.service has finished successfully. ░░ ░░ The job identifier is 236. Dec 14 11:26:33 localhost kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Dec 14 11:26:33 localhost systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-vconsole-setup.service has successfully entered the 'dead' state. Dec 14 11:26:33 localhost systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. ░░ Subject: A stop job for unit systemd-vconsole-setup.service has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit systemd-vconsole-setup.service has finished. ░░ ░░ The job identifier is 300 and the job result is done. Dec 14 11:26:33 localhost systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... ░░ Subject: A start job for unit systemd-vconsole-setup.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-vconsole-setup.service has begun execution. ░░ ░░ The job identifier is 300. Dec 14 11:26:33 localhost systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. ░░ Subject: A start job for unit systemd-vconsole-setup.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-vconsole-setup.service has finished successfully. ░░ ░░ The job identifier is 300. Dec 14 11:26:33 localhost systemd[1]: Reached target sysinit.target - System Initialization. ░░ Subject: A start job for unit sysinit.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sysinit.target has finished successfully. ░░ ░░ The job identifier is 123. Dec 14 11:26:33 localhost systemd[1]: Started dnf-makecache.timer - dnf makecache --timer. ░░ Subject: A start job for unit dnf-makecache.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dnf-makecache.timer has finished successfully. ░░ ░░ The job identifier is 204. Dec 14 11:26:33 localhost systemd[1]: Started fstrim.timer - Discard unused filesystem blocks once a week. ░░ Subject: A start job for unit fstrim.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit fstrim.timer has finished successfully. ░░ ░░ The job identifier is 202. Dec 14 11:26:33 localhost systemd[1]: Started logrotate.timer - Daily rotation of log files. ░░ Subject: A start job for unit logrotate.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.timer has finished successfully. ░░ ░░ The job identifier is 203. Dec 14 11:26:33 localhost systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. ░░ Subject: A start job for unit systemd-tmpfiles-clean.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-tmpfiles-clean.timer has finished successfully. ░░ ░░ The job identifier is 211. Dec 14 11:26:33 localhost systemd[1]: Reached target timers.target - Timer Units. ░░ Subject: A start job for unit timers.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit timers.target has finished successfully. ░░ ░░ The job identifier is 201. Dec 14 11:26:33 localhost systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. ░░ Subject: A start job for unit dbus.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus.socket has finished successfully. ░░ ░░ The job identifier is 208. Dec 14 11:26:33 localhost systemd[1]: Listening on pcscd.socket - PC/SC Smart Card Daemon Activation Socket. ░░ Subject: A start job for unit pcscd.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit pcscd.socket has finished successfully. ░░ ░░ The job identifier is 214. Dec 14 11:26:33 localhost systemd[1]: Listening on sssd-kcm.socket - SSSD Kerberos Cache Manager responder socket. ░░ Subject: A start job for unit sssd-kcm.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sssd-kcm.socket has finished successfully. ░░ ░░ The job identifier is 218. Dec 14 11:26:33 localhost systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. ░░ Subject: A start job for unit systemd-hostnamed.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.socket has finished successfully. ░░ ░░ The job identifier is 221. Dec 14 11:26:33 localhost systemd[1]: Reached target sockets.target - Socket Units. ░░ Subject: A start job for unit sockets.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sockets.target has finished successfully. ░░ ░░ The job identifier is 213. Dec 14 11:26:33 localhost systemd[1]: Starting dbus-broker.service - D-Bus System Message Bus... ░░ Subject: A start job for unit dbus-broker.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus-broker.service has begun execution. ░░ ░░ The job identifier is 209. Dec 14 11:26:33 localhost systemd[1]: systemd-pcrphase-sysinit.service - TPM PCR Barrier (Initialization) was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ░░ Subject: A start job for unit systemd-pcrphase-sysinit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-pcrphase-sysinit.service has finished successfully. ░░ ░░ The job identifier is 198. Dec 14 11:26:33 localhost systemd[1]: Started dbus-broker.service - D-Bus System Message Bus. ░░ Subject: A start job for unit dbus-broker.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dbus-broker.service has finished successfully. ░░ ░░ The job identifier is 209. Dec 14 11:26:33 localhost systemd[1]: Reached target basic.target - Basic System. ░░ Subject: A start job for unit basic.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit basic.target has finished successfully. ░░ ░░ The job identifier is 122. Dec 14 11:26:33 localhost dbus-broker-launch[646]: Ready Dec 14 11:26:33 localhost systemd[1]: Starting chronyd.service - NTP client/server... ░░ Subject: A start job for unit chronyd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has begun execution. ░░ ░░ The job identifier is 256. Dec 14 11:26:33 localhost systemd[1]: Starting cloud-init-local.service - Initial cloud-init job (pre-networking)... ░░ Subject: A start job for unit cloud-init-local.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init-local.service has begun execution. ░░ ░░ The job identifier is 268. Dec 14 11:26:33 localhost systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... ░░ Subject: A start job for unit dracut-shutdown.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dracut-shutdown.service has begun execution. ░░ ░░ The job identifier is 173. Dec 14 11:26:33 localhost systemd[1]: Started irqbalance.service - irqbalance daemon. ░░ Subject: A start job for unit irqbalance.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit irqbalance.service has finished successfully. ░░ ░░ The job identifier is 263. Dec 14 11:26:33 localhost systemd[1]: Started rngd.service - Hardware RNG Entropy Gatherer Daemon. ░░ Subject: A start job for unit rngd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rngd.service has finished successfully. ░░ ░░ The job identifier is 261. Dec 14 11:26:33 localhost (qbalance)[651]: irqbalance.service: Referenced but unset environment variable evaluates to an empty string: IRQBALANCE_ARGS Dec 14 11:26:33 localhost systemd[1]: Starting rsyslog.service - System Logging Service... ░░ Subject: A start job for unit rsyslog.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rsyslog.service has begun execution. ░░ ░░ The job identifier is 259. Dec 14 11:26:33 localhost systemd[1]: ssh-host-keys-migration.service - Update OpenSSH host key permissions was skipped because of an unmet condition check (ConditionPathExists=!/var/lib/.ssh-host-keys-migration). ░░ Subject: A start job for unit ssh-host-keys-migration.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit ssh-host-keys-migration.service has finished successfully. ░░ ░░ The job identifier is 275. Dec 14 11:26:33 localhost systemd[1]: sshd-keygen@ecdsa.service - OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ecdsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ecdsa.service has finished successfully. ░░ ░░ The job identifier is 274. Dec 14 11:26:33 localhost systemd[1]: sshd-keygen@ed25519.service - OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ed25519.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ed25519.service has finished successfully. ░░ ░░ The job identifier is 273. Dec 14 11:26:33 localhost systemd[1]: sshd-keygen@rsa.service - OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@rsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@rsa.service has finished successfully. ░░ ░░ The job identifier is 271. Dec 14 11:26:33 localhost systemd[1]: Reached target sshd-keygen.target. ░░ Subject: A start job for unit sshd-keygen.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen.target has finished successfully. ░░ ░░ The job identifier is 270. Dec 14 11:26:33 localhost systemd[1]: sssd.service - System Security Services Daemon was skipped because no trigger condition checks were met. ░░ Subject: A start job for unit sssd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sssd.service has finished successfully. ░░ ░░ The job identifier is 249. Dec 14 11:26:33 localhost systemd[1]: Reached target nss-user-lookup.target - User and Group Name Lookups. ░░ Subject: A start job for unit nss-user-lookup.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit nss-user-lookup.target has finished successfully. ░░ ░░ The job identifier is 250. Dec 14 11:26:33 localhost systemd[1]: Starting systemd-logind.service - User Login Management... ░░ Subject: A start job for unit systemd-logind.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-logind.service has begun execution. ░░ ░░ The job identifier is 253. Dec 14 11:26:33 localhost systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-credentials-systemd\x2dvconsole\x2dsetup.service.mount has successfully entered the 'dead' state. Dec 14 11:26:33 localhost systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. ░░ Subject: A start job for unit dracut-shutdown.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit dracut-shutdown.service has finished successfully. ░░ ░░ The job identifier is 173. Dec 14 11:26:33 localhost systemd-logind[655]: New seat seat0. ░░ Subject: A new seat seat0 is now available ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new seat seat0 has been configured and is now available. Dec 14 11:26:33 localhost systemd-logind[655]: Watching system buttons on /dev/input/event0 (Power Button) Dec 14 11:26:33 localhost systemd-logind[655]: Watching system buttons on /dev/input/event1 (Sleep Button) Dec 14 11:26:33 localhost systemd-logind[655]: Watching system buttons on /dev/input/event2 (AT Translated Set 2 keyboard) Dec 14 11:26:33 localhost systemd[1]: Started systemd-logind.service - User Login Management. ░░ Subject: A start job for unit systemd-logind.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-logind.service has finished successfully. ░░ ░░ The job identifier is 253. Dec 14 11:26:33 localhost systemd[1]: Started rsyslog.service - System Logging Service. ░░ Subject: A start job for unit rsyslog.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rsyslog.service has finished successfully. ░░ ░░ The job identifier is 259. Dec 14 11:26:33 localhost rsyslogd[654]: imjournal: filecreatemode is not set, using default 0644 [v8.2408.0-2.el10 try https://www.rsyslog.com/e/2186 ] Dec 14 11:26:33 localhost rsyslogd[654]: [origin software="rsyslogd" swVersion="8.2408.0-2.el10" x-pid="654" x-info="https://www.rsyslog.com"] start Dec 14 11:26:33 localhost rsyslogd[654]: imjournal: journal files changed, reloading... [v8.2408.0-2.el10 try https://www.rsyslog.com/e/0 ] Dec 14 11:26:33 localhost chronyd[662]: chronyd version 4.6.1 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +NTS +SECHASH +IPV6 +DEBUG) Dec 14 11:26:33 localhost chronyd[662]: Frequency 0.000 +/- 1000000.000 ppm read from /var/lib/chrony/drift Dec 14 11:26:33 localhost chronyd[662]: Loaded seccomp filter (level 2) Dec 14 11:26:33 localhost systemd[1]: Started chronyd.service - NTP client/server. ░░ Subject: A start job for unit chronyd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit chronyd.service has finished successfully. ░░ ░░ The job identifier is 256. Dec 14 11:26:33 localhost rngd[653]: Disabling 7: PKCS11 Entropy generator (pkcs11) Dec 14 11:26:33 localhost rngd[653]: Disabling 5: NIST Network Entropy Beacon (nist) Dec 14 11:26:33 localhost rngd[653]: Disabling 9: Qrypt quantum entropy beacon (qrypt) Dec 14 11:26:33 localhost rngd[653]: Disabling 10: Named pipe entropy input (namedpipe) Dec 14 11:26:33 localhost rngd[653]: Initializing available sources Dec 14 11:26:33 localhost rngd[653]: [hwrng ]: Initialization Failed Dec 14 11:26:33 localhost rngd[653]: [rdrand]: Enabling RDRAND rng support Dec 14 11:26:33 localhost rngd[653]: [rdrand]: Initialized Dec 14 11:26:33 localhost rngd[653]: [jitter]: JITTER timeout set to 5 sec Dec 14 11:26:33 localhost rngd[653]: [jitter]: Initializing AES buffer Dec 14 11:26:37 localhost cloud-init[669]: Cloud-init v. 24.1.4-21.el10 running 'init-local' at Sat, 14 Dec 2024 16:26:37 +0000. Up 18.36 seconds. Dec 14 11:26:37 localhost dhcpcd[671]: dhcpcd-10.0.6 starting Dec 14 11:26:37 localhost kernel: 8021q: 802.1Q VLAN Support v1.8 Dec 14 11:26:37 localhost systemd[1]: Listening on systemd-rfkill.socket - Load/Save RF Kill Switch Status /dev/rfkill Watch. ░░ Subject: A start job for unit systemd-rfkill.socket has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-rfkill.socket has finished successfully. ░░ ░░ The job identifier is 310. Dec 14 11:26:37 localhost kernel: cfg80211: Loading compiled-in X.509 certificates for regulatory database Dec 14 11:26:37 localhost kernel: Loaded X.509 cert 'sforshee: 00b28ddf47aef9cea7' Dec 14 11:26:37 localhost kernel: Loaded X.509 cert 'wens: 61c038651aabdcf94bd0ac7ff06c7248db18c600' Dec 14 11:26:37 localhost dhcpcd[674]: DUID 00:01:00:01:2e:f0:6e:3d:0a:ff:fb:3d:09:35 Dec 14 11:26:37 localhost dhcpcd[674]: eth0: IAID fb:3d:09:35 Dec 14 11:26:37 localhost kernel: platform regulatory.0: Direct firmware load for regulatory.db failed with error -2 Dec 14 11:26:37 localhost kernel: cfg80211: failed to load regulatory.db Dec 14 11:26:38 localhost rngd[653]: [jitter]: Unable to obtain AES key, disabling JITTER source Dec 14 11:26:38 localhost rngd[653]: [jitter]: Initialization Failed Dec 14 11:26:38 localhost rngd[653]: Process privileges have been dropped to 2:2 Dec 14 11:26:38 localhost dhcpcd[674]: eth0: soliciting a DHCP lease Dec 14 11:26:38 localhost dhcpcd[674]: eth0: offered 10.31.14.110 from 10.31.12.1 Dec 14 11:26:38 localhost dhcpcd[674]: eth0: leased 10.31.14.110 for 3600 seconds Dec 14 11:26:38 localhost dhcpcd[674]: eth0: adding route to 10.31.12.0/22 Dec 14 11:26:38 localhost dhcpcd[674]: eth0: adding default route via 10.31.12.1 Dec 14 11:26:38 localhost dhcpcd[674]: control command: /usr/sbin/dhcpcd --dumplease --ipv4only eth0 Dec 14 11:26:38 localhost systemd[1]: Starting systemd-hostnamed.service - Hostname Service... ░░ Subject: A start job for unit systemd-hostnamed.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has begun execution. ░░ ░░ The job identifier is 319. Dec 14 11:26:38 localhost systemd[1]: Started systemd-hostnamed.service - Hostname Service. ░░ Subject: A start job for unit systemd-hostnamed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has finished successfully. ░░ ░░ The job identifier is 319. Dec 14 11:26:38 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-hostnamed[696]: Hostname set to (static) Dec 14 11:26:38 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished cloud-init-local.service - Initial cloud-init job (pre-networking). ░░ Subject: A start job for unit cloud-init-local.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init-local.service has finished successfully. ░░ ░░ The job identifier is 268. Dec 14 11:26:38 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target network-pre.target - Preparation for Network. ░░ Subject: A start job for unit network-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network-pre.target has finished successfully. ░░ ░░ The job identifier is 159. Dec 14 11:26:38 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting NetworkManager.service - Network Manager... ░░ Subject: A start job for unit NetworkManager.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager.service has begun execution. ░░ ░░ The job identifier is 207. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.2519] NetworkManager (version 1.51.4-1.el10) is starting... (boot:67ab4bf4-984b-4f4d-8c19-30dde3975820) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.2521] Read config: /etc/NetworkManager/NetworkManager.conf, /etc/NetworkManager/conf.d/30-cloud-init-ip6-addr-gen-mode.conf Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3090] manager[0x5614ecd3f9f0]: monitoring kernel firmware directory '/lib/firmware'. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3122] hostname: hostname: using hostnamed Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3122] hostname: static hostname changed from (none) to "ip-10-31-14-110.us-east-1.aws.redhat.com" Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3129] dns-mgr: init: dns=default,systemd-resolved rc-manager=symlink (auto) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3133] manager[0x5614ecd3f9f0]: rfkill: Wi-Fi hardware radio set enabled Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3134] manager[0x5614ecd3f9f0]: rfkill: WWAN hardware radio set enabled Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3181] manager: rfkill: Wi-Fi enabled by radio killswitch; enabled by state file Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3182] manager: rfkill: WWAN enabled by radio killswitch; enabled by state file Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3182] manager: Networking is enabled by state file Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3213] settings: Loaded settings plugin: keyfile (internal) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 397. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3538] dhcp: init: Using DHCP client 'internal' Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3540] manager: (lo): new Loopback device (/org/freedesktop/NetworkManager/Devices/1) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3549] device (lo): state change: unmanaged -> unavailable (reason 'connection-assumed', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3566] device (lo): state change: unavailable -> disconnected (reason 'connection-assumed', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3572] device (lo): Activation: starting connection 'lo' (8e242600-3605-43e7-b815-8160476017ad) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3578] manager: (eth0): new Ethernet device (/org/freedesktop/NetworkManager/Devices/2) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3581] device (eth0): state change: unmanaged -> unavailable (reason 'managed', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started NetworkManager.service - Network Manager. ░░ Subject: A start job for unit NetworkManager.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager.service has finished successfully. ░░ ░░ The job identifier is 207. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3596] bus-manager: acquired D-Bus service "org.freedesktop.NetworkManager" Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3599] device (lo): state change: disconnected -> prepare (reason 'none', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3610] device (lo): state change: prepare -> config (reason 'none', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3612] device (lo): state change: config -> ip-config (reason 'none', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3621] device (eth0): carrier: link connected Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target network.target - Network. ░░ Subject: A start job for unit network.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network.target has finished successfully. ░░ ░░ The job identifier is 210. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3641] device (lo): state change: ip-config -> ip-check (reason 'none', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3646] device (eth0): state change: unavailable -> disconnected (reason 'carrier-changed', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3650] policy: auto-activating connection 'cloud-init eth0' (1dd9a779-d327-56e1-8454-c65e2556c12c) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3654] device (eth0): Activation: starting connection 'cloud-init eth0' (1dd9a779-d327-56e1-8454-c65e2556c12c) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3655] device (eth0): state change: disconnected -> prepare (reason 'none', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3657] manager: NetworkManager state is now CONNECTING Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3658] device (eth0): state change: prepare -> config (reason 'none', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3662] device (eth0): state change: config -> ip-config (reason 'none', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3671] dhcp4 (eth0): activation: beginning transaction (timeout in 45 seconds) Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.3689] dhcp4 (eth0): state changed new lease, address=10.31.14.110, acd pending Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting NetworkManager-wait-online.service - Network Manager Wait Online... ░░ Subject: A start job for unit NetworkManager-wait-online.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-wait-online.service has begun execution. ░░ ░░ The job identifier is 206. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting gssproxy.service - GSSAPI Proxy Daemon... ░░ Subject: A start job for unit gssproxy.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit gssproxy.service has begun execution. ░░ ░░ The job identifier is 244. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 397. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5143] dhcp4 (eth0): state changed new lease, address=10.31.14.110 Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5181] policy: set 'cloud-init eth0' (eth0) as default for IPv4 routing and DNS Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5904] device (lo): state change: ip-check -> secondaries (reason 'none', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5909] device (eth0): state change: ip-config -> ip-check (reason 'none', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5918] device (lo): state change: secondaries -> activated (reason 'none', managed-type: 'external') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5922] device (lo): Activation: successful, device activated. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5970] device (eth0): state change: ip-check -> secondaries (reason 'none', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5976] device (eth0): state change: secondaries -> activated (reason 'none', managed-type: 'full') Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5980] manager: NetworkManager state is now CONNECTED_SITE Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5983] device (eth0): Activation: successful, device activated. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5987] manager: NetworkManager state is now CONNECTED_GLOBAL Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com NetworkManager[703]: [1734193599.5990] manager: startup complete Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished NetworkManager-wait-online.service - Network Manager Wait Online. ░░ Subject: A start job for unit NetworkManager-wait-online.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-wait-online.service has finished successfully. ░░ ░░ The job identifier is 206. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting cloud-init.service - Initial cloud-init job (metadata service crawler)... ░░ Subject: A start job for unit cloud-init.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.service has begun execution. ░░ ░░ The job identifier is 267. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started gssproxy.service - GSSAPI Proxy Daemon. ░░ Subject: A start job for unit gssproxy.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit gssproxy.service has finished successfully. ░░ ░░ The job identifier is 244. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: rpc-gssd.service - RPC security service for NFS client and server was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). ░░ Subject: A start job for unit rpc-gssd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-gssd.service has finished successfully. ░░ ░░ The job identifier is 240. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target nfs-client.target - NFS client services. ░░ Subject: A start job for unit nfs-client.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit nfs-client.target has finished successfully. ░░ ░░ The job identifier is 238. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. ░░ Subject: A start job for unit remote-fs-pre.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-fs-pre.target has finished successfully. ░░ ░░ The job identifier is 245. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. ░░ Subject: A start job for unit remote-cryptsetup.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-cryptsetup.target has finished successfully. ░░ ░░ The job identifier is 234. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target remote-fs.target - Remote File Systems. ░░ Subject: A start job for unit remote-fs.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit remote-fs.target has finished successfully. ░░ ░░ The job identifier is 277. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: systemd-pcrphase.service - TPM PCR Barrier (User) was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ░░ Subject: A start job for unit systemd-pcrphase.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-pcrphase.service has finished successfully. ░░ ░░ The job identifier is 196. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com chronyd[662]: Added source 10.11.160.238 Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com chronyd[662]: Added source 10.18.100.10 Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com chronyd[662]: Added source 10.2.32.37 Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com chronyd[662]: Added source 10.2.32.38 Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Cloud-init v. 24.1.4-21.el10 running 'init' at Sat, 14 Dec 2024 16:26:39 +0000. Up 21.09 seconds. Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: ++++++++++++++++++++++++++++++++++++++Net device info++++++++++++++++++++++++++++++++++++++ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +--------+------+----------------------------+---------------+--------+-------------------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | Device | Up | Address | Mask | Scope | Hw-Address | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +--------+------+----------------------------+---------------+--------+-------------------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | eth0 | True | 10.31.14.110 | 255.255.252.0 | global | 0a:ff:fb:3d:09:35 | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | eth0 | True | fe80::8ff:fbff:fe3d:935/64 | . | link | 0a:ff:fb:3d:09:35 | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | lo | True | 127.0.0.1 | 255.0.0.0 | host | . | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | lo | True | ::1/128 | . | host | . | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +--------+------+----------------------------+---------------+--------+-------------------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: ++++++++++++++++++++++++++++Route IPv4 info+++++++++++++++++++++++++++++ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +-------+-------------+------------+---------------+-----------+-------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | Route | Destination | Gateway | Genmask | Interface | Flags | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +-------+-------------+------------+---------------+-----------+-------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | 0 | 0.0.0.0 | 10.31.12.1 | 0.0.0.0 | eth0 | UG | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | 1 | 10.31.12.0 | 0.0.0.0 | 255.255.252.0 | eth0 | U | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +-------+-------------+------------+---------------+-----------+-------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +++++++++++++++++++Route IPv6 info+++++++++++++++++++ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +-------+-------------+---------+-----------+-------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | Route | Destination | Gateway | Interface | Flags | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +-------+-------------+---------+-----------+-------+ Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | 0 | fe80::/64 | :: | eth0 | U | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: | 2 | multicast | :: | eth0 | U | Dec 14 11:26:39 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: ci-info: +-------+-------------+---------+-----------+-------+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Generating public/private rsa key pair. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Your identification has been saved in /etc/ssh/ssh_host_rsa_key Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Your public key has been saved in /etc/ssh/ssh_host_rsa_key.pub Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: The key fingerprint is: Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: SHA256:6zYh8D9sJVTMc/UplqhDb1JRoJSmKvg9yVrtJKOf8Y0 root@ip-10-31-14-110.us-east-1.aws.redhat.com Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: The key's randomart image is: Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: +---[RSA 3072]----+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | +.ooo. | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | .o* + ...| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | o+ = + ..| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . .o + . . | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . o..S+ o | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . . .+ o.= | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . +=.*.+ | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | .o=O.X | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | ooo.E.+ | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: +----[SHA256]-----+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Generating public/private ecdsa key pair. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Your identification has been saved in /etc/ssh/ssh_host_ecdsa_key Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Your public key has been saved in /etc/ssh/ssh_host_ecdsa_key.pub Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: The key fingerprint is: Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: SHA256:2/VGYFFBaPbV9mIKV6Gxe7Knrqzyong5LsvYtXW93UY root@ip-10-31-14-110.us-east-1.aws.redhat.com Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: The key's randomart image is: Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: +---[ECDSA 256]---+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | o+=o.| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | +=. +| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | o=o o.| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | ...o+ .| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | S o+oo. | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | + ..*E | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | ... o o ..+ | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | +.o+oo. . o =. | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: |. ==+o +o.+o+.. | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: +----[SHA256]-----+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Generating public/private ed25519 key pair. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Your identification has been saved in /etc/ssh/ssh_host_ed25519_key Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: Your public key has been saved in /etc/ssh/ssh_host_ed25519_key.pub Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: The key fingerprint is: Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: SHA256:nR3w4spyXF/iKNEXvkwO+019RpKTGbhfuvuA8uYmPwU root@ip-10-31-14-110.us-east-1.aws.redhat.com Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: The key's randomart image is: Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: +--[ED25519 256]--+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | o . | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . = . | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | + =E+ = | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | S B B.B o| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | o + @ *.B | Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | . * + B.= +| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | o ..++o +.| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: | *=.+o.| Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[788]: +----[SHA256]-----+ Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished cloud-init.service - Initial cloud-init job (metadata service crawler). ░░ Subject: A start job for unit cloud-init.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.service has finished successfully. ░░ ░░ The job identifier is 267. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target cloud-config.target - Cloud-config availability. ░░ Subject: A start job for unit cloud-config.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.target has finished successfully. ░░ ░░ The job identifier is 266. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target network-online.target - Network is Online. ░░ Subject: A start job for unit network-online.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit network-online.target has finished successfully. ░░ ░░ The job identifier is 205. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting cloud-config.service - Apply the settings specified in cloud-config... ░░ Subject: A start job for unit cloud-config.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.service has begun execution. ░░ ░░ The job identifier is 265. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting kdump.service - Crash recovery kernel arming... ░░ Subject: A start job for unit kdump.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit kdump.service has begun execution. ░░ ░░ The job identifier is 237. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting restraintd.service - The restraint harness.... ░░ Subject: A start job for unit restraintd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit restraintd.service has begun execution. ░░ ░░ The job identifier is 278. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting rpc-statd-notify.service - Notify NFS peers of a restart... ░░ Subject: A start job for unit rpc-statd-notify.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-statd-notify.service has begun execution. ░░ ░░ The job identifier is 246. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting sshd.service - OpenSSH server daemon... ░░ Subject: A start job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 269. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com sm-notify[876]: Version 2.7.1 starting Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started restraintd.service - The restraint harness.. ░░ Subject: A start job for unit restraintd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit restraintd.service has finished successfully. ░░ ░░ The job identifier is 278. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com (sshd)[879]: sshd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started rpc-statd-notify.service - Notify NFS peers of a restart. ░░ Subject: A start job for unit rpc-statd-notify.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit rpc-statd-notify.service has finished successfully. ░░ ░░ The job identifier is 246. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com sshd[879]: Server listening on 0.0.0.0 port 22. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com sshd[879]: Server listening on :: port 22. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started sshd.service - OpenSSH server daemon. ░░ Subject: A start job for unit sshd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has finished successfully. ░░ ░░ The job identifier is 269. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[920]: Cloud-init v. 24.1.4-21.el10 running 'modules:config' at Sat, 14 Dec 2024 16:26:41 +0000. Up 22.75 seconds. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Stopping sshd.service - OpenSSH server daemon... ░░ Subject: A stop job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 501. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com sshd[879]: Received signal 15; terminating. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: sshd.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit sshd.service has successfully entered the 'dead' state. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Stopped sshd.service - OpenSSH server daemon. ░░ Subject: A stop job for unit sshd.service has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd.service has finished. ░░ ░░ The job identifier is 501 and the job result is done. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Stopped target sshd-keygen.target. ░░ Subject: A stop job for unit sshd-keygen.target has finished ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd-keygen.target has finished. ░░ ░░ The job identifier is 585 and the job result is done. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Stopping sshd-keygen.target... ░░ Subject: A stop job for unit sshd-keygen.target has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A stop job for unit sshd-keygen.target has begun execution. ░░ ░░ The job identifier is 585. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: ssh-host-keys-migration.service - Update OpenSSH host key permissions was skipped because of an unmet condition check (ConditionPathExists=!/var/lib/.ssh-host-keys-migration). ░░ Subject: A start job for unit ssh-host-keys-migration.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit ssh-host-keys-migration.service has finished successfully. ░░ ░░ The job identifier is 584. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: sshd-keygen@ecdsa.service - OpenSSH ecdsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ecdsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ecdsa.service has finished successfully. ░░ ░░ The job identifier is 583. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: sshd-keygen@ed25519.service - OpenSSH ed25519 Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@ed25519.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@ed25519.service has finished successfully. ░░ ░░ The job identifier is 582. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: sshd-keygen@rsa.service - OpenSSH rsa Server Key Generation was skipped because of an unmet condition check (ConditionPathExists=!/run/systemd/generator.early/multi-user.target.wants/cloud-init.target). ░░ Subject: A start job for unit sshd-keygen@rsa.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen@rsa.service has finished successfully. ░░ ░░ The job identifier is 580. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target sshd-keygen.target. ░░ Subject: A start job for unit sshd-keygen.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd-keygen.target has finished successfully. ░░ ░░ The job identifier is 585. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting sshd.service - OpenSSH server daemon... ░░ Subject: A start job for unit sshd.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has begun execution. ░░ ░░ The job identifier is 501. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com (sshd)[924]: sshd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com sshd[924]: Server listening on 0.0.0.0 port 22. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com sshd[924]: Server listening on :: port 22. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started sshd.service - OpenSSH server daemon. ░░ Subject: A start job for unit sshd.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit sshd.service has finished successfully. ░░ ░░ The job identifier is 501. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished cloud-config.service - Apply the settings specified in cloud-config. ░░ Subject: A start job for unit cloud-config.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-config.service has finished successfully. ░░ ░░ The job identifier is 265. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting cloud-final.service - Execute cloud user/final scripts... ░░ Subject: A start job for unit cloud-final.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-final.service has begun execution. ░░ ░░ The job identifier is 276. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... ░░ Subject: A start job for unit systemd-user-sessions.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-user-sessions.service has begun execution. ░░ ░░ The job identifier is 262. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. ░░ Subject: A start job for unit systemd-user-sessions.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-user-sessions.service has finished successfully. ░░ ░░ The job identifier is 262. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started crond.service - Command Scheduler. ░░ Subject: A start job for unit crond.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit crond.service has finished successfully. ░░ ░░ The job identifier is 260. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started getty@tty1.service - Getty on tty1. ░░ Subject: A start job for unit getty@tty1.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit getty@tty1.service has finished successfully. ░░ ░░ The job identifier is 232. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. ░░ Subject: A start job for unit serial-getty@ttyS0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit serial-getty@ttyS0.service has finished successfully. ░░ ░░ The job identifier is 227. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target getty.target - Login Prompts. ░░ Subject: A start job for unit getty.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit getty.target has finished successfully. ░░ ░░ The job identifier is 226. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target multi-user.target - Multi-User System. ░░ Subject: A start job for unit multi-user.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit multi-user.target has finished successfully. ░░ ░░ The job identifier is 121. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting systemd-update-utmp-runlevel.service - Record Runlevel Change in UTMP... ░░ Subject: A start job for unit systemd-update-utmp-runlevel.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp-runlevel.service has begun execution. ░░ ░░ The job identifier is 235. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com restraintd[883]: Listening on http://localhost:8081 Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com crond[930]: (CRON) STARTUP (1.7.0) Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com crond[930]: (CRON) INFO (Syslog will be used instead of sendmail.) Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com crond[930]: (CRON) INFO (RANDOM_DELAY will be scaled with factor 7% if used.) Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com crond[930]: (CRON) INFO (running with inotify support) Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-update-utmp-runlevel.service has successfully entered the 'dead' state. Dec 14 11:26:41 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished systemd-update-utmp-runlevel.service - Record Runlevel Change in UTMP. ░░ Subject: A start job for unit systemd-update-utmp-runlevel.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-update-utmp-runlevel.service has finished successfully. ░░ ░░ The job identifier is 235. Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com kdumpctl[881]: kdump: Detected change(s) in the following file(s): /etc/fstab Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1035]: Cloud-init v. 24.1.4-21.el10 running 'modules:final' at Sat, 14 Dec 2024 16:26:42 +0000. Up 23.23 seconds. Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1080]: ############################################################# Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1081]: -----BEGIN SSH HOST KEY FINGERPRINTS----- Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1083]: 256 SHA256:2/VGYFFBaPbV9mIKV6Gxe7Knrqzyong5LsvYtXW93UY root@ip-10-31-14-110.us-east-1.aws.redhat.com (ECDSA) Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1088]: 256 SHA256:nR3w4spyXF/iKNEXvkwO+019RpKTGbhfuvuA8uYmPwU root@ip-10-31-14-110.us-east-1.aws.redhat.com (ED25519) Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1094]: 3072 SHA256:6zYh8D9sJVTMc/UplqhDb1JRoJSmKvg9yVrtJKOf8Y0 root@ip-10-31-14-110.us-east-1.aws.redhat.com (RSA) Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1098]: -----END SSH HOST KEY FINGERPRINTS----- Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1099]: ############################################################# Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com cloud-init[1035]: Cloud-init v. 24.1.4-21.el10 finished at Sat, 14 Dec 2024 16:26:42 +0000. Datasource DataSourceEc2Local. Up 23.40 seconds Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished cloud-final.service - Execute cloud user/final scripts. ░░ Subject: A start job for unit cloud-final.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-final.service has finished successfully. ░░ ░░ The job identifier is 276. Dec 14 11:26:42 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Reached target cloud-init.target - Cloud-init target. ░░ Subject: A start job for unit cloud-init.target has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit cloud-init.target has finished successfully. ░░ ░░ The job identifier is 264. Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 0 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 0 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 48 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 48 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 49 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 49 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 50 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 50 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 51 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 51 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 52 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 52 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 53 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 53 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 54 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 54 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 55 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 55 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 56 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 56 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 57 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 57 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 58 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 58 affinity is now unmanaged Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: Cannot change IRQ 59 affinity: Permission denied Dec 14 11:26:43 ip-10-31-14-110.us-east-1.aws.redhat.com irqbalance[651]: IRQ 59 affinity is now unmanaged Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com kernel: block xvda: the capability attribute has been deprecated. Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com kdumpctl[881]: kdump: Rebuilding /boot/initramfs-6.12.0-31.el10.x86_64kdump.img Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1388]: dracut-103-1.el10 Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1391]: Executing: /usr/bin/dracut --list-modules Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1462]: dracut-103-1.el10 Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Executing: /usr/bin/dracut --add kdumpbase --quiet --hostonly --hostonly-cmdline --hostonly-i18n --hostonly-mode strict --hostonly-nics --aggressive-strip --omit "rdma plymouth resume ifcfg earlykdump" --mount "/dev/disk/by-uuid/f3bb1e80-fac3-4b5e-93f6-d763469176c6 /sysroot xfs rw,relatime,seclabel,attr2,inode64,logbufs=8,logbsize=32k,noquota" --add squash-squashfs --squash-compressor zstd --no-hostonly-default-device -f /boot/initramfs-6.12.0-31.el10.x86_64kdump.img 6.12.0-31.el10.x86_64 Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-bsod' will not be installed, because command '/usr/lib/systemd/systemd-bsod' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-networkd' will not be installed, because command '/usr/lib/systemd/systemd-networkd' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-networkd' will not be installed, because command '/usr/lib/systemd/systemd-networkd-wait-online' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-pcrphase' will not be installed, because command '/usr/lib/systemd/systemd-pcrphase' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-portabled' will not be installed, because command 'portablectl' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-portabled' will not be installed, because command '/usr/lib/systemd/systemd-portabled' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-resolved' will not be installed, because command '/usr/lib/systemd/systemd-resolved' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-timesyncd' could not be found! Dec 14 11:26:45 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-time-wait-sync' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'busybox' will not be installed, because command 'busybox' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com chronyd[662]: Selected source 10.2.32.38 Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'dbus-daemon' will not be installed, because command 'dbus-daemon' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'connman' will not be installed, because command 'connmand' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'connman' will not be installed, because command 'connmanctl' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'connman' will not be installed, because command 'connmand-wait-online' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'ifcfg' will not be installed, because it's in the list to be omitted! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'plymouth' will not be installed, because it's in the list to be omitted! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: 62bluetooth: Could not find any command of '/usr/lib/bluetooth/bluetoothd /usr/libexec/bluetooth/bluetoothd'! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'btrfs' will not be installed, because command 'btrfs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'dmraid' will not be installed, because command 'dmraid' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'mdraid' will not be installed, because command 'mdadm' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'multipath' will not be installed, because command 'multipath' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'iscsi' will not be installed, because command 'iscsid' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'nvmf' will not be installed, because command 'nvme' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'resume' will not be installed, because it's in the list to be omitted! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'squash-erofs' will not be installed, because command 'mkfs.erofs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'squash-erofs' will not be installed, because command 'fsck.erofs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'biosdevname' will not be installed, because command 'biosdevname' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'earlykdump' will not be installed, because it's in the list to be omitted! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-bsod' will not be installed, because command '/usr/lib/systemd/systemd-bsod' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-pcrphase' will not be installed, because command '/usr/lib/systemd/systemd-pcrphase' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-portabled' will not be installed, because command 'portablectl' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-portabled' will not be installed, because command '/usr/lib/systemd/systemd-portabled' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-resolved' will not be installed, because command '/usr/lib/systemd/systemd-resolved' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-timesyncd' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'systemd-timesyncd' will not be installed, because command '/usr/lib/systemd/systemd-time-wait-sync' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'busybox' will not be installed, because command 'busybox' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'dbus-daemon' will not be installed, because command 'dbus-daemon' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'connman' will not be installed, because command 'connmand' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'connman' will not be installed, because command 'connmanctl' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'connman' will not be installed, because command 'connmand-wait-online' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: 62bluetooth: Could not find any command of '/usr/lib/bluetooth/bluetoothd /usr/libexec/bluetooth/bluetoothd'! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'btrfs' will not be installed, because command 'btrfs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'dmraid' will not be installed, because command 'dmraid' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'mdraid' will not be installed, because command 'mdadm' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'multipath' will not be installed, because command 'multipath' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'cifs' will not be installed, because command 'mount.cifs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'iscsi' will not be installed, because command 'iscsi-iname' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'iscsi' will not be installed, because command 'iscsiadm' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'iscsi' will not be installed, because command 'iscsid' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'nvmf' will not be installed, because command 'nvme' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'squash-erofs' will not be installed, because command 'mkfs.erofs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Module 'squash-erofs' will not be installed, because command 'fsck.erofs' could not be found! Dec 14 11:26:46 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: fips *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: fips-crypto-policies *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-ask-password *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-initrd *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-journald *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-modules-load *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-sysctl *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-sysusers *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-tmpfiles *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: systemd-udevd *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: rngd *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: i18n *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: drm *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: prefixdevname *** Dec 14 11:26:47 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: kernel-modules *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: kernel-modules-extra *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: kernel-modules-extra: configuration source "/run/depmod.d" does not exist Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: kernel-modules-extra: configuration source "/lib/depmod.d" does not exist Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: kernel-modules-extra: parsing configuration file "/etc/depmod.d/dist.conf" Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: kernel-modules-extra: /etc/depmod.d/dist.conf: added "updates extra built-in weak-updates" to the list of search directories Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: pcmcia *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Skipping udev rule: 60-pcmcia.rules Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: fstab-sys *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: hwdb *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: rootfs-block *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: squash-squashfs *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: terminfo *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: udev-rules *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: dracut-systemd *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: usrmount *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: base *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: fs-lib *** Dec 14 11:26:48 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: kdumpbase *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: memstrack *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: microcode_ctl-fw_dir_override *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl module: mangling fw_dir Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl: reset fw_dir to "/lib/firmware/updates /lib/firmware" Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel"... Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl: intel: caveats check for kernel version "6.12.0-31.el10.x86_64" passed, adding "/usr/share/microcode_ctl/ucode_with_caveats/intel" to fw_dir variable Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl: processing data directory "/usr/share/microcode_ctl/ucode_with_caveats/intel-06-4f-01"... Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl: configuration "intel-06-4f-01" is ignored Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: microcode_ctl: final fw_dir: "/usr/share/microcode_ctl/ucode_with_caveats/intel /lib/firmware/updates /lib/firmware" Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: shutdown *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including module: squash-lib *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Including modules done *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Installing kernel module dependencies *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Installing kernel module dependencies done *** Dec 14 11:26:49 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Resolving executable dependencies *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Resolving executable dependencies done *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Hardlinking files *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Mode: real Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Method: sha256 Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Files: 537 Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Linked: 25 files Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Compared: 0 xattrs Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Compared: 48 files Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Saved: 13.58 MiB Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Duration: 0.152061 seconds Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Hardlinking files done *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Generating early-microcode cpio image *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Constructing GenuineIntel.bin *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Constructing GenuineIntel.bin *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Store current command line parameters *** Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: Stored kernel commandline: Dec 14 11:26:50 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: No dracut internal kernel commandline stored in the initramfs Dec 14 11:26:51 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Squashing the files inside the initramfs *** Dec 14 11:26:58 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Squashing the files inside the initramfs done *** Dec 14 11:26:58 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Creating image file '/boot/initramfs-6.12.0-31.el10.x86_64kdump.img' *** Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com dracut[1465]: *** Creating initramfs image file '/boot/initramfs-6.12.0-31.el10.x86_64kdump.img' done *** Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com kernel: PKCS7: Message signed outside of X.509 validity window Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com kdumpctl[881]: kdump: kexec: loaded kdump kernel Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com kdumpctl[881]: kdump: Starting kdump: [OK] Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com kdumpctl[881]: kdump: Notice: No vmcore creation test performed! Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished kdump.service - Crash recovery kernel arming. ░░ Subject: A start job for unit kdump.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit kdump.service has finished successfully. ░░ ░░ The job identifier is 237. Dec 14 11:26:59 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Startup finished in 995ms (kernel) + 5.404s (initrd) + 34.706s (userspace) = 41.106s. ░░ Subject: System start-up is now complete ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ All system services necessary queued for starting at boot have been ░░ started. Note that this does not mean that the machine is now idle as services ░░ might still be busy with completing start-up. ░░ ░░ Kernel start-up required 995066 microseconds. ░░ ░░ Initrd start-up required 5404966 microseconds. ░░ ░░ Userspace start-up required 34706427 microseconds. Dec 14 11:27:09 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: systemd-hostnamed.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-hostnamed.service has successfully entered the 'dead' state. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4326]: Accepted publickey for root from 10.30.34.106 port 56984 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4326]: pam_systemd(sshd:session): New sd-bus connection (system-bus-pam-systemd-4326) opened. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Created slice user-0.slice - User Slice of UID 0. ░░ Subject: A start job for unit user-0.slice has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-0.slice has finished successfully. ░░ ░░ The job identifier is 673. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting user-runtime-dir@0.service - User Runtime Directory /run/user/0... ░░ Subject: A start job for unit user-runtime-dir@0.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-runtime-dir@0.service has begun execution. ░░ ░░ The job identifier is 595. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: New session 1 of user root. ░░ Subject: A new session 1 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 1 has been created for the user root. ░░ ░░ The leading process of the session is 4326. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Finished user-runtime-dir@0.service - User Runtime Directory /run/user/0. ░░ Subject: A start job for unit user-runtime-dir@0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user-runtime-dir@0.service has finished successfully. ░░ ░░ The job identifier is 595. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting user@0.service - User Manager for UID 0... ░░ Subject: A start job for unit user@0.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user@0.service has begun execution. ░░ ░░ The job identifier is 675. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: New session 2 of user root. ░░ Subject: A new session 2 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 2 has been created for the user root. ░░ ░░ The leading process of the session is 4331. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com (systemd)[4331]: pam_unix(systemd-user:session): session opened for user root(uid=0) by root(uid=0) Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Queued start job for default target default.target. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Created slice app.slice - User Application Slice. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 5. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: grub-boot-success.timer - Mark boot as successful after the user session has run 2 minutes was skipped because of an unmet condition check (ConditionUser=!@system). ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 11. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of User's Temporary Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 10. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Reached target paths.target - Paths. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 8. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Reached target timers.target - Timers. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 9. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Starting dbus.socket - D-Bus User Message Bus Socket... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 4. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Starting systemd-tmpfiles-setup.service - Create User Files and Directories... ░░ Subject: A start job for unit UNIT has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has begun execution. ░░ ░░ The job identifier is 12. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Finished systemd-tmpfiles-setup.service - Create User Files and Directories. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 12. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Listening on dbus.socket - D-Bus User Message Bus Socket. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 4. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Reached target sockets.target - Sockets. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 3. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Reached target basic.target - Basic System. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 2. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Reached target default.target - Main User Target. ░░ Subject: A start job for unit UNIT has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit UNIT has finished successfully. ░░ ░░ The job identifier is 1. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[4331]: Startup finished in 112ms. ░░ Subject: User manager start-up is now complete ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The user manager instance for user 0 has been started. All services queued ░░ for starting have been started. Note that other services might still be starting ░░ up or be started at any later time. ░░ ░░ Startup of the manager took 112474 microseconds. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started user@0.service - User Manager for UID 0. ░░ Subject: A start job for unit user@0.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit user@0.service has finished successfully. ░░ ░░ The job identifier is 675. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started session-1.scope - Session 1 of User root. ░░ Subject: A start job for unit session-1.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-1.scope has finished successfully. ░░ ░░ The job identifier is 756. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4326]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4342]: Received disconnect from 10.30.34.106 port 56984:11: disconnected by user Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4342]: Disconnected from user root 10.30.34.106 port 56984 Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4326]: pam_systemd(sshd:session): New sd-bus connection (system-bus-pam-systemd-4326) opened. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4326]: pam_unix(sshd:session): session closed for user root Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: session-1.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-1.scope has successfully entered the 'dead' state. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: Session 1 logged out. Waiting for processes to exit. Dec 14 11:29:20 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: Removed session 1. ░░ Subject: Session 1 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 1 has been terminated. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4380]: Accepted publickey for root from 10.31.8.152 port 46578 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4381]: Accepted publickey for root from 10.31.8.152 port 46592 ssh2: RSA SHA256:W3cSdmPJK+d9RwU97ardijPXIZnxHswrpTHWW9oYtEU Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4380]: pam_systemd(sshd:session): New sd-bus connection (system-bus-pam-systemd-4380) opened. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4381]: pam_systemd(sshd:session): New sd-bus connection (system-bus-pam-systemd-4381) opened. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: New session 3 of user root. ░░ Subject: A new session 3 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 3 has been created for the user root. ░░ ░░ The leading process of the session is 4380. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started session-3.scope - Session 3 of User root. ░░ Subject: A start job for unit session-3.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-3.scope has finished successfully. ░░ ░░ The job identifier is 838. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: New session 4 of user root. ░░ Subject: A new session 4 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 4 has been created for the user root. ░░ ░░ The leading process of the session is 4381. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started session-4.scope - Session 4 of User root. ░░ Subject: A start job for unit session-4.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-4.scope has finished successfully. ░░ ░░ The job identifier is 920. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4380]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4381]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4387]: Received disconnect from 10.31.8.152 port 46592:11: disconnected by user Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4387]: Disconnected from user root 10.31.8.152 port 46592 Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4381]: pam_systemd(sshd:session): New sd-bus connection (system-bus-pam-systemd-4381) opened. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com sshd-session[4381]: pam_unix(sshd:session): session closed for user root Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: Session 4 logged out. Waiting for processes to exit. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: session-4.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-4.scope has successfully entered the 'dead' state. Dec 14 11:29:23 ip-10-31-14-110.us-east-1.aws.redhat.com systemd-logind[655]: Removed session 4. ░░ Subject: Session 4 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 4 has been terminated. Dec 14 11:29:28 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Starting systemd-hostnamed.service - Hostname Service... ░░ Subject: A start job for unit systemd-hostnamed.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has begun execution. ░░ ░░ The job identifier is 1002. Dec 14 11:29:28 ip-10-31-14-110.us-east-1.aws.redhat.com systemd[1]: Started systemd-hostnamed.service - Hostname Service. ░░ Subject: A start job for unit systemd-hostnamed.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit systemd-hostnamed.service has finished successfully. ░░ ░░ The job identifier is 1002. Dec 14 11:29:28 managed-node3 systemd-hostnamed[5861]: Hostname set to (static) Dec 14 11:29:28 managed-node3 NetworkManager[703]: [1734193768.6417] hostname: static hostname changed from "ip-10-31-14-110.us-east-1.aws.redhat.com" to "managed-node3" Dec 14 11:29:28 managed-node3 systemd[1]: Starting NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 1080. Dec 14 11:29:28 managed-node3 systemd[1]: Started NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 1080. Dec 14 11:29:38 managed-node3 systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Dec 14 11:29:58 managed-node3 systemd[1]: systemd-hostnamed.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit systemd-hostnamed.service has successfully entered the 'dead' state. Dec 14 11:30:08 managed-node3 sshd-session[6527]: Accepted publickey for root from 10.31.13.174 port 49050 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Dec 14 11:30:08 managed-node3 sshd-session[6527]: pam_systemd(sshd:session): New sd-bus connection (system-bus-pam-systemd-6527) opened. Dec 14 11:30:08 managed-node3 systemd-logind[655]: New session 5 of user root. ░░ Subject: A new session 5 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 5 has been created for the user root. ░░ ░░ The leading process of the session is 6527. Dec 14 11:30:08 managed-node3 systemd[1]: Started session-5.scope - Session 5 of User root. ░░ Subject: A start job for unit session-5.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-5.scope has finished successfully. ░░ ░░ The job identifier is 1159. Dec 14 11:30:08 managed-node3 sshd-session[6527]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Dec 14 11:30:10 managed-node3 python3.12[6683]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Dec 14 11:30:11 managed-node3 python3.12[6843]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:30:11 managed-node3 python3.12[6974]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:30:13 managed-node3 sudo[7236]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-xclcopowrlvzimvlwodbfevotsixbavv ; /usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1734193813.0266075-7127-254553333473826/AnsiballZ_dnf.py' Dec 14 11:30:13 managed-node3 sudo[7236]: pam_systemd(sudo:session): New sd-bus connection (system-bus-pam-systemd-7236) opened. Dec 14 11:30:13 managed-node3 sudo[7236]: pam_unix(sudo:session): session opened for user root(uid=0) by root(uid=0) Dec 14 11:30:13 managed-node3 python3.12[7239]: ansible-ansible.legacy.dnf Invoked with name=['iptables-nft', 'podman', 'shadow-utils-subid'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Dec 14 11:30:27 managed-node3 kernel: SELinux: Converting 384 SID table entries... Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability network_peer_controls=1 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability open_perms=1 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability extended_socket_class=1 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability always_check_network=0 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 14 11:30:27 managed-node3 kernel: SELinux: policy capability userspace_initial_context=0 Dec 14 11:30:35 managed-node3 kernel: SELinux: Converting 385 SID table entries... Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability network_peer_controls=1 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability open_perms=1 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability extended_socket_class=1 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability always_check_network=0 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 14 11:30:35 managed-node3 kernel: SELinux: policy capability userspace_initial_context=0 Dec 14 11:30:43 managed-node3 kernel: SELinux: Converting 385 SID table entries... Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability network_peer_controls=1 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability open_perms=1 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability extended_socket_class=1 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability always_check_network=0 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 14 11:30:43 managed-node3 kernel: SELinux: policy capability userspace_initial_context=0 Dec 14 11:30:44 managed-node3 setsebool[7313]: The virt_use_nfs policy boolean was changed to 1 by root Dec 14 11:30:44 managed-node3 setsebool[7313]: The virt_sandbox_use_all_caps policy boolean was changed to 1 by root Dec 14 11:30:54 managed-node3 kernel: SELinux: Converting 388 SID table entries... Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability network_peer_controls=1 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability open_perms=1 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability extended_socket_class=1 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability always_check_network=0 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 14 11:30:54 managed-node3 kernel: SELinux: policy capability userspace_initial_context=0 Dec 14 11:31:02 managed-node3 kernel: SELinux: Converting 388 SID table entries... Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability network_peer_controls=1 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability open_perms=1 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability extended_socket_class=1 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability always_check_network=0 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability genfs_seclabel_symlinks=1 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 14 11:31:02 managed-node3 kernel: SELinux: policy capability userspace_initial_context=0 Dec 14 11:31:19 managed-node3 systemd[1]: Started run-r716e48b3056747a5bf3b0cc91f601af1.service - /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r716e48b3056747a5bf3b0cc91f601af1.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r716e48b3056747a5bf3b0cc91f601af1.service has finished successfully. ░░ ░░ The job identifier is 1241. Dec 14 11:31:19 managed-node3 systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1319. Dec 14 11:31:19 managed-node3 systemd[1]: Reload requested from client PID 8038 ('systemctl') (unit session-5.scope)... Dec 14 11:31:19 managed-node3 systemd[1]: Reloading... Dec 14 11:31:19 managed-node3 systemd[1]: Reloading finished in 254 ms. Dec 14 11:31:19 managed-node3 systemd[1]: Queuing reload/restart jobs for marked units… Dec 14 11:31:20 managed-node3 sudo[7236]: pam_unix(sudo:session): session closed for user root Dec 14 11:31:21 managed-node3 python3.12[8493]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:21 managed-node3 systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Dec 14 11:31:21 managed-node3 systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1319. Dec 14 11:31:21 managed-node3 systemd[1]: run-r716e48b3056747a5bf3b0cc91f601af1.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-r716e48b3056747a5bf3b0cc91f601af1.service has successfully entered the 'dead' state. Dec 14 11:31:22 managed-node3 python3.12[8634]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Dec 14 11:31:22 managed-node3 python3.12[8766]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:24 managed-node3 python3.12[8899]: ansible-tempfile Invoked with state=directory prefix=lsr_podman_config_ suffix= path=None Dec 14 11:31:24 managed-node3 python3.12[9030]: ansible-ansible.legacy.command Invoked with _raw_params=tar --ignore-failed-read -c -P -v -p -f /tmp/lsr_podman_config_umo1wg40/backup.tar /etc/containers/containers.conf.d/50-systemroles.conf /etc/containers/registries.conf.d/50-systemroles.conf /etc/containers/storage.conf /etc/containers/policy.json _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:25 managed-node3 python3.12[9162]: ansible-user Invoked with name=user1 state=present non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on managed-node3 update_password=always uid=None group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None password_expire_warn=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None Dec 14 11:31:25 managed-node3 useradd[9164]: new group: name=user1, GID=1000 Dec 14 11:31:25 managed-node3 useradd[9164]: new user: name=user1, UID=1000, GID=1000, home=/home/user1, shell=/bin/bash, from=/dev/pts/0 Dec 14 11:31:27 managed-node3 python3.12[9426]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:28 managed-node3 python3.12[9564]: ansible-getent Invoked with database=passwd key=user1 fail_key=False service=None split=None Dec 14 11:31:29 managed-node3 python3.12[9696]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:29 managed-node3 python3.12[9829]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:29 managed-node3 python3.12[9961]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:30 managed-node3 python3.12[10093]: ansible-file Invoked with path=/home/user1/.config/containers/containers.conf.d state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:31 managed-node3 python3.12[10224]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:31 managed-node3 python3.12[10329]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734193891.0325742-8177-94420310335864/.source.conf dest=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf owner=user1 mode=0644 follow=False _original_basename=toml.j2 checksum=94370d6e765779f1c58daf02f667b8f0b74d91f6 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:32 managed-node3 python3.12[10460]: ansible-file Invoked with path=/home/user1/.config/containers/registries.conf.d state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:32 managed-node3 python3.12[10591]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:32 managed-node3 python3.12[10696]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734193892.3603973-8232-1526005629345/.source.conf dest=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf owner=user1 mode=0644 follow=False _original_basename=toml.j2 checksum=dfb9cd7094a81b3d1bb06512cc9b49a09c75639b backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:33 managed-node3 python3.12[10827]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:34 managed-node3 python3.12[10958]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:34 managed-node3 python3.12[11063]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734193893.732546-8296-55796330500548/.source.conf dest=/home/user1/.config/containers/storage.conf owner=user1 mode=0644 follow=False _original_basename=toml.j2 checksum=d08574b6a1df63dbe1c939ff0bcc7c0b61d03044 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:34 managed-node3 python3.12[11194]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:35 managed-node3 python3.12[11325]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:35 managed-node3 python3.12[11456]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:36 managed-node3 python3.12[11561]: ansible-ansible.legacy.copy Invoked with dest=/home/user1/.config/containers/policy.json owner=user1 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1734193895.4946458-8411-208121650010930/.source.json _original_basename=.tddnhefz follow=False checksum=6746c079ad563b735fc39f73d4876654b80b0a0d backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:36 managed-node3 python3.12[11692]: ansible-stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:37 managed-node3 python3.12[11825]: ansible-stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:37 managed-node3 python3.12[11958]: ansible-stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:37 managed-node3 python3.12[12091]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:39 managed-node3 python3.12[12355]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:40 managed-node3 python3.12[12492]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:41 managed-node3 python3.12[12625]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:41 managed-node3 python3.12[12757]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:42 managed-node3 python3.12[12889]: ansible-file Invoked with path=/home/user1/.config/containers/containers.conf.d state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:42 managed-node3 python3.12[13020]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:42 managed-node3 python3.12[13086]: ansible-ansible.legacy.file Invoked with owner=user1 mode=0644 dest=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf _original_basename=toml.j2 recurse=False state=file path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:43 managed-node3 python3.12[13217]: ansible-file Invoked with path=/home/user1/.config/containers/registries.conf.d state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:43 managed-node3 python3.12[13348]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:44 managed-node3 python3.12[13414]: ansible-ansible.legacy.file Invoked with owner=user1 mode=0644 dest=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf _original_basename=toml.j2 recurse=False state=file path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:44 managed-node3 python3.12[13545]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:45 managed-node3 python3.12[13676]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:45 managed-node3 python3.12[13742]: ansible-ansible.legacy.file Invoked with owner=user1 mode=0644 dest=/home/user1/.config/containers/storage.conf _original_basename=toml.j2 recurse=False state=file path=/home/user1/.config/containers/storage.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:45 managed-node3 python3.12[13873]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:46 managed-node3 python3.12[14004]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:46 managed-node3 python3.12[14137]: ansible-slurp Invoked with path=/home/user1/.config/containers/policy.json src=/home/user1/.config/containers/policy.json Dec 14 11:31:47 managed-node3 python3.12[14268]: ansible-stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:48 managed-node3 python3.12[14401]: ansible-stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:48 managed-node3 python3.12[14534]: ansible-stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:48 managed-node3 python3.12[14667]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:51 managed-node3 python3.12[14931]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:31:52 managed-node3 python3.12[15069]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Dec 14 11:31:52 managed-node3 python3.12[15201]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:54 managed-node3 python3.12[15334]: ansible-file Invoked with path=/etc/containers/containers.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:54 managed-node3 python3.12[15465]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:55 managed-node3 python3.12[15570]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734193914.33953-9204-179873555081338/.source.conf dest=/etc/containers/containers.conf.d/50-systemroles.conf owner=root mode=0644 follow=False _original_basename=toml.j2 checksum=94370d6e765779f1c58daf02f667b8f0b74d91f6 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:55 managed-node3 python3.12[15701]: ansible-file Invoked with path=/etc/containers/registries.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:56 managed-node3 python3.12[15832]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:56 managed-node3 python3.12[15937]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734193915.975161-9276-112706284746257/.source.conf dest=/etc/containers/registries.conf.d/50-systemroles.conf owner=root mode=0644 follow=False _original_basename=toml.j2 checksum=dfb9cd7094a81b3d1bb06512cc9b49a09c75639b backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:57 managed-node3 python3.12[16068]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:57 managed-node3 python3.12[16199]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:31:58 managed-node3 python3.12[16304]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734193917.5309792-9356-138504811101551/.source.conf dest=/etc/containers/storage.conf owner=root mode=0644 follow=False _original_basename=toml.j2 checksum=d08574b6a1df63dbe1c939ff0bcc7c0b61d03044 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:58 managed-node3 python3.12[16435]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:31:59 managed-node3 python3.12[16566]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:31:59 managed-node3 python3.12[16699]: ansible-slurp Invoked with path=/etc/containers/policy.json src=/etc/containers/policy.json Dec 14 11:32:00 managed-node3 python3.12[16830]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:32:00 managed-node3 python3.12[16937]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/policy.json owner=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1734193919.7730489-9479-127190054442142/.source.json _original_basename=.d8o5zfsi follow=False checksum=6746c079ad563b735fc39f73d4876654b80b0a0d backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:01 managed-node3 python3.12[17068]: ansible-stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:02 managed-node3 python3.12[17201]: ansible-stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:02 managed-node3 python3.12[17334]: ansible-stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:02 managed-node3 python3.12[17467]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:05 managed-node3 python3.12[17731]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:32:06 managed-node3 python3.12[17868]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:07 managed-node3 python3.12[18001]: ansible-file Invoked with path=/etc/containers/containers.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:08 managed-node3 python3.12[18132]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:32:08 managed-node3 python3.12[18198]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/containers.conf.d/50-systemroles.conf _original_basename=toml.j2 recurse=False state=file path=/etc/containers/containers.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:09 managed-node3 python3.12[18329]: ansible-file Invoked with path=/etc/containers/registries.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:09 managed-node3 python3.12[18460]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:32:09 managed-node3 python3.12[18526]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/registries.conf.d/50-systemroles.conf _original_basename=toml.j2 recurse=False state=file path=/etc/containers/registries.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:10 managed-node3 python3.12[18657]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:11 managed-node3 python3.12[18788]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:32:11 managed-node3 python3.12[18854]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/storage.conf _original_basename=toml.j2 recurse=False state=file path=/etc/containers/storage.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:11 managed-node3 python3.12[18985]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:12 managed-node3 python3.12[19116]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:12 managed-node3 python3.12[19249]: ansible-slurp Invoked with path=/etc/containers/policy.json src=/etc/containers/policy.json Dec 14 11:32:14 managed-node3 python3.12[19380]: ansible-stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:14 managed-node3 python3.12[19513]: ansible-stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:14 managed-node3 python3.12[19646]: ansible-stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:15 managed-node3 python3.12[19779]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:16 managed-node3 python3.12[19912]: ansible-slurp Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf src=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf Dec 14 11:32:16 managed-node3 python3.12[20043]: ansible-slurp Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf src=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf Dec 14 11:32:17 managed-node3 python3.12[20174]: ansible-slurp Invoked with path=/home/user1/.config/containers/storage.conf src=/home/user1/.config/containers/storage.conf Dec 14 11:32:17 managed-node3 python3.12[20305]: ansible-slurp Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf src=/etc/containers/containers.conf.d/50-systemroles.conf Dec 14 11:32:18 managed-node3 python3.12[20436]: ansible-slurp Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf src=/etc/containers/registries.conf.d/50-systemroles.conf Dec 14 11:32:19 managed-node3 python3.12[20567]: ansible-slurp Invoked with path=/etc/containers/storage.conf src=/etc/containers/storage.conf Dec 14 11:32:19 managed-node3 python3.12[20698]: ansible-file Invoked with state=absent path=/etc/containers/containers.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:20 managed-node3 python3.12[20829]: ansible-file Invoked with state=absent path=/etc/containers/registries.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:20 managed-node3 python3.12[20960]: ansible-file Invoked with state=absent path=/etc/containers/storage.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:21 managed-node3 python3.12[21091]: ansible-file Invoked with state=absent path=/etc/containers/policy.json recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:21 managed-node3 python3.12[21222]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:21 managed-node3 python3.12[21353]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:22 managed-node3 python3.12[21484]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/storage.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:22 managed-node3 python3.12[21615]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/policy.json recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:23 managed-node3 python3.12[21746]: ansible-ansible.legacy.command Invoked with _raw_params=tar xfvpP /tmp/lsr_podman_config_umo1wg40/backup.tar _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:32:23 managed-node3 python3.12[21878]: ansible-file Invoked with state=absent path=/tmp/lsr_podman_config_umo1wg40 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:32:26 managed-node3 python3.12[22052]: ansible-setup Invoked with gather_subset=['!all', '!min', 'distribution', 'distribution_major_version', 'distribution_version', 'os_family'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Dec 14 11:32:26 managed-node3 python3.12[22185]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:27 managed-node3 python3.12[22316]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:29 managed-node3 python3.12[22578]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:32:30 managed-node3 python3.12[22715]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Dec 14 11:32:31 managed-node3 python3.12[22847]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:35 managed-node3 python3.12[23023]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Dec 14 11:32:40 managed-node3 python3.12[23183]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:40 managed-node3 python3.12[23314]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:43 managed-node3 python3.12[23576]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:32:44 managed-node3 python3.12[23714]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Dec 14 11:32:45 managed-node3 python3.12[23846]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:51 managed-node3 python3.12[24022]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Dec 14 11:32:53 managed-node3 python3.12[24182]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:32:53 managed-node3 python3.12[24313]: ansible-ansible.legacy.dnf Invoked with name=['python3-pyasn1', 'python3-cryptography', 'python3-dbus'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Dec 14 11:32:56 managed-node3 python3.12[24449]: ansible-ansible.legacy.dnf Invoked with name=['certmonger', 'python3-packaging'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Dec 14 11:32:57 managed-node3 dbus-broker-launch[646]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Dec 14 11:32:57 managed-node3 dbus-broker-launch[646]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Dec 14 11:32:57 managed-node3 dbus-broker-launch[646]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Dec 14 11:32:57 managed-node3 systemd[1]: Reload requested from client PID 24457 ('systemctl') (unit session-5.scope)... Dec 14 11:32:57 managed-node3 systemd[1]: Reloading... Dec 14 11:32:58 managed-node3 systemd[1]: Reloading finished in 185 ms. Dec 14 11:32:58 managed-node3 systemd[1]: Started run-re25853d457974692b318b9a879e60844.service - /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-re25853d457974692b318b9a879e60844.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-re25853d457974692b318b9a879e60844.service has finished successfully. ░░ ░░ The job identifier is 1401. Dec 14 11:32:58 managed-node3 systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1479. Dec 14 11:32:58 managed-node3 systemd[1]: Reload requested from client PID 24519 ('systemctl') (unit session-5.scope)... Dec 14 11:32:58 managed-node3 systemd[1]: Reloading... Dec 14 11:32:58 managed-node3 systemd[1]: Reloading finished in 291 ms. Dec 14 11:32:58 managed-node3 systemd[1]: Queuing reload/restart jobs for marked units… Dec 14 11:32:59 managed-node3 systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Dec 14 11:32:59 managed-node3 systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1479. Dec 14 11:32:59 managed-node3 systemd[1]: run-re25853d457974692b318b9a879e60844.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-re25853d457974692b318b9a879e60844.service has successfully entered the 'dead' state. Dec 14 11:32:59 managed-node3 python3.12[24715]: ansible-file Invoked with name=/etc/certmonger//pre-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//pre-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:00 managed-node3 python3.12[24846]: ansible-file Invoked with name=/etc/certmonger//post-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//post-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:00 managed-node3 python3.12[24977]: ansible-ansible.legacy.systemd Invoked with name=certmonger state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Dec 14 11:33:00 managed-node3 systemd[1]: Reload requested from client PID 24980 ('systemctl') (unit session-5.scope)... Dec 14 11:33:00 managed-node3 systemd[1]: Reloading... Dec 14 11:33:01 managed-node3 systemd[1]: Reloading finished in 181 ms. Dec 14 11:33:01 managed-node3 systemd[1]: Starting certmonger.service - Certificate monitoring and PKI enrollment... ░░ Subject: A start job for unit certmonger.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit certmonger.service has begun execution. ░░ ░░ The job identifier is 1557. Dec 14 11:33:01 managed-node3 (rtmonger)[25034]: certmonger.service: Referenced but unset environment variable evaluates to an empty string: OPTS Dec 14 11:33:01 managed-node3 systemd[1]: Started certmonger.service - Certificate monitoring and PKI enrollment. ░░ Subject: A start job for unit certmonger.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit certmonger.service has finished successfully. ░░ ░░ The job identifier is 1557. Dec 14 11:33:01 managed-node3 python3.12[25192]: ansible-fedora.linux_system_roles.certificate_request Invoked with name=quadlet_demo dns=['localhost'] directory=/etc/pki/tls wait=True ca=self-sign __header=# # Ansible managed # # system_role:certificate provider_config_directory=/etc/certmonger provider=certmonger key_usage=['digitalSignature', 'keyEncipherment'] extended_key_usage=['id-kp-serverAuth', 'id-kp-clientAuth'] auto_renew=True ip=None email=None common_name=None country=None state=None locality=None organization=None organizational_unit=None contact_email=None key_size=None owner=None group=None mode=None principal=None run_before=None run_after=None Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:01 managed-node3 certmonger[25034]: 2024-12-14 11:33:01 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 certmonger[25207]: Certificate in file "/etc/pki/tls/certs/quadlet_demo.crt" issued by CA and saved. Dec 14 11:33:02 managed-node3 certmonger[25034]: 2024-12-14 11:33:02 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:02 managed-node3 python3.12[25338]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Dec 14 11:33:03 managed-node3 python3.12[25469]: ansible-slurp Invoked with path=/etc/pki/tls/private/quadlet_demo.key src=/etc/pki/tls/private/quadlet_demo.key Dec 14 11:33:03 managed-node3 python3.12[25600]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Dec 14 11:33:04 managed-node3 python3.12[25731]: ansible-ansible.legacy.command Invoked with _raw_params=getcert stop-tracking -f /etc/pki/tls/certs/quadlet_demo.crt _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:33:04 managed-node3 certmonger[25034]: 2024-12-14 11:33:04 [25034] Wrote to /var/lib/certmonger/requests/20241214163301 Dec 14 11:33:04 managed-node3 python3.12[25863]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:05 managed-node3 python3.12[25994]: ansible-file Invoked with path=/etc/pki/tls/private/quadlet_demo.key state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:05 managed-node3 python3.12[26125]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:06 managed-node3 python3.12[26256]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:06 managed-node3 python3.12[26387]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:08 managed-node3 python3.12[26649]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:33:10 managed-node3 python3.12[26787]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Dec 14 11:33:10 managed-node3 python3.12[26919]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:13 managed-node3 python3.12[27052]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:14 managed-node3 python3.12[27183]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:14 managed-node3 python3.12[27314]: ansible-ansible.legacy.dnf Invoked with name=['firewalld'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Dec 14 11:33:16 managed-node3 python3.12[27446]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False scope=system no_block=False state=None enabled=None force=None Dec 14 11:33:16 managed-node3 python3.12[27579]: ansible-ansible.legacy.systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Dec 14 11:33:16 managed-node3 systemd[1]: Reload requested from client PID 27582 ('systemctl') (unit session-5.scope)... Dec 14 11:33:16 managed-node3 systemd[1]: Reloading... Dec 14 11:33:16 managed-node3 systemd[1]: Reloading finished in 185 ms. Dec 14 11:33:17 managed-node3 systemd[1]: Starting firewalld.service - firewalld - dynamic firewall daemon... ░░ Subject: A start job for unit firewalld.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit firewalld.service has begun execution. ░░ ░░ The job identifier is 1636. Dec 14 11:33:17 managed-node3 systemd[1]: Started firewalld.service - firewalld - dynamic firewall daemon. ░░ Subject: A start job for unit firewalld.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit firewalld.service has finished successfully. ░░ ░░ The job identifier is 1636. Dec 14 11:33:17 managed-node3 systemd[1]: Starting polkit.service - Authorization Manager... ░░ Subject: A start job for unit polkit.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit polkit.service has begun execution. ░░ ░░ The job identifier is 1720. Dec 14 11:33:17 managed-node3 polkitd[27747]: Started polkitd version 125 Dec 14 11:33:17 managed-node3 systemd[1]: Started polkit.service - Authorization Manager. ░░ Subject: A start job for unit polkit.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit polkit.service has finished successfully. ░░ ░░ The job identifier is 1720. Dec 14 11:33:18 managed-node3 python3.12[27820]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Dec 14 11:33:18 managed-node3 python3.12[27951]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Dec 14 11:33:20 managed-node3 systemd[1]: var-lib-containers-storage-overlay-compat279581914-merged.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay-compat279581914-merged.mount has successfully entered the 'dead' state. Dec 14 11:33:20 managed-node3 kernel: evm: overlay not supported Dec 14 11:33:20 managed-node3 podman[28090]: 2024-12-14 11:33:20.380826987 -0500 EST m=+0.086593393 system refresh Dec 14 11:33:21 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:21 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:21 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:22 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:22 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:24 managed-node3 python3.12[28560]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:25 managed-node3 python3.12[28693]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:26 managed-node3 python3.12[28824]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo.network follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:33:26 managed-node3 python3.12[28929]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734194006.0269196-14057-133971458264709/.source.network dest=/etc/containers/systemd/quadlet-demo.network owner=root group=0 mode=0644 _original_basename=quadlet-demo.network follow=False checksum=e57c08d49aff4bae8daab138d913aeddaa8682a0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:27 managed-node3 python3.12[29060]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:33:27 managed-node3 systemd[1]: Reload requested from client PID 29061 ('systemctl') (unit session-5.scope)... Dec 14 11:33:27 managed-node3 systemd[1]: Reloading... Dec 14 11:33:27 managed-node3 systemd[1]: Reloading finished in 189 ms. Dec 14 11:33:28 managed-node3 python3.12[29244]: ansible-systemd Invoked with name=quadlet-demo-network.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Dec 14 11:33:28 managed-node3 systemd[1]: Starting quadlet-demo-network.service... ░░ Subject: A start job for unit quadlet-demo-network.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-network.service has begun execution. ░░ ░░ The job identifier is 1800. Dec 14 11:33:28 managed-node3 quadlet-demo-network[29248]: time="2024-12-14T11:33:28-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" Dec 14 11:33:28 managed-node3 quadlet-demo-network[29248]: systemd-quadlet-demo Dec 14 11:33:28 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:28 managed-node3 systemd[1]: Finished quadlet-demo-network.service. ░░ Subject: A start job for unit quadlet-demo-network.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-network.service has finished successfully. ░░ ░░ The job identifier is 1800. Dec 14 11:33:29 managed-node3 python3.12[29387]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:30 managed-node3 python3.12[29520]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:30 managed-node3 python3.12[29651]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:33:31 managed-node3 python3.12[29756]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734194010.6736438-14236-92540342954756/.source.volume dest=/etc/containers/systemd/quadlet-demo-mysql.volume owner=root group=0 mode=0644 _original_basename=quadlet-demo-mysql.volume follow=False checksum=585f8cbdf0ec73000f9227dcffbef71e9552ea4a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:31 managed-node3 python3.12[29887]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:33:31 managed-node3 systemd[1]: Reload requested from client PID 29888 ('systemctl') (unit session-5.scope)... Dec 14 11:33:31 managed-node3 systemd[1]: Reloading... Dec 14 11:33:32 managed-node3 systemd[1]: Reloading finished in 187 ms. Dec 14 11:33:32 managed-node3 python3.12[30071]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Dec 14 11:33:32 managed-node3 systemd[1]: Starting quadlet-demo-mysql-volume.service... ░░ Subject: A start job for unit quadlet-demo-mysql-volume.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql-volume.service has begun execution. ░░ ░░ The job identifier is 1884. Dec 14 11:33:32 managed-node3 quadlet-demo-mysql-volume[30075]: time="2024-12-14T11:33:32-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" Dec 14 11:33:32 managed-node3 podman[30075]: 2024-12-14 11:33:32.664121916 -0500 EST m=+0.030237627 volume create systemd-quadlet-demo-mysql Dec 14 11:33:32 managed-node3 quadlet-demo-mysql-volume[30075]: systemd-quadlet-demo-mysql Dec 14 11:33:32 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:32 managed-node3 systemd[1]: Finished quadlet-demo-mysql-volume.service. ░░ Subject: A start job for unit quadlet-demo-mysql-volume.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql-volume.service has finished successfully. ░░ ░░ The job identifier is 1884. Dec 14 11:33:33 managed-node3 python3.12[30214]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:35 managed-node3 python3.12[30347]: ansible-file Invoked with path=/tmp/quadlet_demo state=directory owner=root group=root mode=0777 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:35 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:38 managed-node3 systemd[1]: var-lib-containers-storage-overlay-compat424278761-lower\x2dmapped.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay-compat424278761-lower\x2dmapped.mount has successfully entered the 'dead' state. Dec 14 11:33:42 managed-node3 podman[30487]: 2024-12-14 11:33:42.478789036 -0500 EST m=+6.634414537 image pull dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6 Dec 14 11:33:42 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:42 managed-node3 rsyslogd[654]: imjournal: journal files changed, reloading... [v8.2408.0-2.el10 try https://www.rsyslog.com/e/0 ] Dec 14 11:33:42 managed-node3 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Dec 14 11:33:42 managed-node3 python3.12[30806]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:43 managed-node3 python3.12[30937]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:33:43 managed-node3 python3.12[31042]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/systemd/quadlet-demo-mysql.container owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1734194023.0743153-14590-125136684732491/.source.container _original_basename=.be5euwc5 follow=False checksum=ca62b2ad3cc9afb5b5371ebbf797b9bc4fd7edd4 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:44 managed-node3 python3.12[31173]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:33:44 managed-node3 systemd[1]: Reload requested from client PID 31174 ('systemctl') (unit session-5.scope)... Dec 14 11:33:44 managed-node3 systemd[1]: Reloading... Dec 14 11:33:44 managed-node3 systemd[1]: Reloading finished in 189 ms. Dec 14 11:33:44 managed-node3 python3.12[31357]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Dec 14 11:33:44 managed-node3 systemd[1]: Starting quadlet-demo-mysql.service... ░░ Subject: A start job for unit quadlet-demo-mysql.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql.service has begun execution. ░░ ░░ The job identifier is 1968. Dec 14 11:33:45 managed-node3 quadlet-demo-mysql[31361]: time="2024-12-14T11:33:45-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" Dec 14 11:33:45 managed-node3 podman[31361]: 2024-12-14 11:33:45.059579002 -0500 EST m=+0.044945043 container create e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Dec 14 11:33:45 managed-node3 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1085] manager: (podman1): new Bridge device (/org/freedesktop/NetworkManager/Devices/3) Dec 14 11:33:45 managed-node3 kernel: podman1: port 1(veth0) entered blocking state Dec 14 11:33:45 managed-node3 kernel: podman1: port 1(veth0) entered disabled state Dec 14 11:33:45 managed-node3 kernel: veth0: entered allmulticast mode Dec 14 11:33:45 managed-node3 kernel: veth0: entered promiscuous mode Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1254] manager: (veth0): new Veth device (/org/freedesktop/NetworkManager/Devices/4) Dec 14 11:33:45 managed-node3 kernel: podman1: port 1(veth0) entered blocking state Dec 14 11:33:45 managed-node3 kernel: podman1: port 1(veth0) entered forwarding state Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1290] device (veth0): carrier: link connected Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1292] device (podman1): carrier: link connected Dec 14 11:33:45 managed-node3 podman[31361]: 2024-12-14 11:33:45.043155252 -0500 EST m=+0.028521433 image pull dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 quay.io/linux-system-roles/mysql:5.6 Dec 14 11:33:45 managed-node3 (udev-worker)[31378]: Network interface NamePolicy= disabled on kernel command line. Dec 14 11:33:45 managed-node3 (udev-worker)[31380]: Network interface NamePolicy= disabled on kernel command line. Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1732] device (podman1): state change: unmanaged -> unavailable (reason 'connection-assumed', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1741] device (podman1): state change: unavailable -> disconnected (reason 'connection-assumed', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1748] device (podman1): Activation: starting connection 'podman1' (b4057425-e67a-485d-8a7a-63f5d24af0ff) Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1751] device (podman1): state change: disconnected -> prepare (reason 'none', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1754] device (podman1): state change: prepare -> config (reason 'none', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1756] device (podman1): state change: config -> ip-config (reason 'none', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.1759] device (podman1): state change: ip-config -> ip-check (reason 'none', managed-type: 'external') Dec 14 11:33:45 managed-node3 systemd[1]: Starting NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service... ░░ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has begun execution. ░░ ░░ The job identifier is 2055. Dec 14 11:33:45 managed-node3 systemd[1]: Started NetworkManager-dispatcher.service - Network Manager Script Dispatcher Service. ░░ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit NetworkManager-dispatcher.service has finished successfully. ░░ ░░ The job identifier is 2055. Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.2268] device (podman1): state change: ip-check -> secondaries (reason 'none', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.2272] device (podman1): state change: secondaries -> activated (reason 'none', managed-type: 'external') Dec 14 11:33:45 managed-node3 NetworkManager[703]: [1734194025.2282] device (podman1): Activation: successful, device activated. Dec 14 11:33:45 managed-node3 systemd[1]: Started run-r7fbb015e72674f0ba85c639c3ed71a6f.scope - /usr/libexec/podman/aardvark-dns --config /run/containers/networks/aardvark-dns -p 53 run. ░░ Subject: A start job for unit run-r7fbb015e72674f0ba85c639c3ed71a6f.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r7fbb015e72674f0ba85c639c3ed71a6f.scope has finished successfully. ░░ ░░ The job identifier is 2134. Dec 14 11:33:45 managed-node3 systemd[1]: Started e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72-7b63df59ce7fdf47.timer - /usr/bin/podman healthcheck run e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72. ░░ Subject: A start job for unit e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72-7b63df59ce7fdf47.timer has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72-7b63df59ce7fdf47.timer has finished successfully. ░░ ░░ The job identifier is 2140. Dec 14 11:33:45 managed-node3 podman[31361]: 2024-12-14 11:33:45.31271762 -0500 EST m=+0.298083770 container init e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Dec 14 11:33:45 managed-node3 systemd[1]: Started quadlet-demo-mysql.service. ░░ Subject: A start job for unit quadlet-demo-mysql.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo-mysql.service has finished successfully. ░░ ░░ The job identifier is 1968. Dec 14 11:33:45 managed-node3 podman[31361]: 2024-12-14 11:33:45.338985245 -0500 EST m=+0.324351383 container start e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Dec 14 11:33:45 managed-node3 quadlet-demo-mysql[31361]: e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 Dec 14 11:33:45 managed-node3 podman[31429]: 2024-12-14 11:33:45.496049336 -0500 EST m=+0.120347512 container health_status e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, health_status=healthy, health_failing_streak=0, health_log=, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Dec 14 11:33:46 managed-node3 python3.12[31625]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:47 managed-node3 python3.12[31769]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:48 managed-node3 python3.12[31900]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:33:48 managed-node3 python3.12[32005]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734194027.8082154-14774-224715775555281/.source.yml dest=/etc/containers/systemd/envoy-proxy-configmap.yml owner=root group=0 mode=0644 _original_basename=envoy-proxy-configmap.yml follow=False checksum=d681c7d56f912150d041873e880818b22a90c188 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:49 managed-node3 python3.12[32160]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:33:49 managed-node3 systemd[1]: Reload requested from client PID 32161 ('systemctl') (unit session-5.scope)... Dec 14 11:33:49 managed-node3 systemd[1]: Reloading... Dec 14 11:33:49 managed-node3 systemd[1]: Reloading finished in 206 ms. Dec 14 11:33:49 managed-node3 systemd[1]: Starting logrotate.service - Rotate log files... ░░ Subject: A start job for unit logrotate.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.service has begun execution. ░░ ░░ The job identifier is 2296. Dec 14 11:33:49 managed-node3 systemd[1]: logrotate.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit logrotate.service has successfully entered the 'dead' state. Dec 14 11:33:49 managed-node3 systemd[1]: Finished logrotate.service - Rotate log files. ░░ Subject: A start job for unit logrotate.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit logrotate.service has finished successfully. ░░ ░░ The job identifier is 2296. Dec 14 11:33:50 managed-node3 python3.12[32349]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:52 managed-node3 python3.12[32519]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:52 managed-node3 python3.12[32650]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo.yml follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:33:53 managed-node3 python3.12[32755]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/systemd/quadlet-demo.yml owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1734194032.2620435-14941-256399581694956/.source.yml _original_basename=.mz1rbt_9 follow=False checksum=998dccde0483b1654327a46ddd89cbaa47650370 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:53 managed-node3 python3.12[32893]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:33:53 managed-node3 systemd[1]: Reload requested from client PID 32894 ('systemctl') (unit session-5.scope)... Dec 14 11:33:53 managed-node3 systemd[1]: Reloading... Dec 14 11:33:53 managed-node3 systemd[1]: Reloading finished in 214 ms. Dec 14 11:33:54 managed-node3 python3.12[33078]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:33:55 managed-node3 systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Dec 14 11:33:55 managed-node3 python3.12[33235]: ansible-slurp Invoked with path=/etc/containers/systemd/quadlet-demo.yml src=/etc/containers/systemd/quadlet-demo.yml Dec 14 11:33:56 managed-node3 python3.12[33366]: ansible-file Invoked with path=/tmp/httpd3 state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:33:57 managed-node3 python3.12[33497]: ansible-file Invoked with path=/tmp/httpd3-create state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:34:09 managed-node3 podman[33636]: 2024-12-14 11:34:09.806030447 -0500 EST m=+12.237925783 image pull fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b quay.io/linux-system-roles/wordpress:4.8-apache Dec 14 11:34:15 managed-node3 podman[34066]: 2024-12-14 11:34:15.170385157 -0500 EST m=+4.909223490 image pull 5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d quay.io/linux-system-roles/envoyproxy:v1.25.0 Dec 14 11:34:15 managed-node3 python3.12[34337]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:34:15 managed-node3 podman[34446]: 2024-12-14 11:34:15.95217748 -0500 EST m=+0.115361640 container health_status e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, health_status=healthy, health_failing_streak=0, health_log=, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Dec 14 11:34:16 managed-node3 python3.12[34474]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/quadlet-demo.kube follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Dec 14 11:34:16 managed-node3 python3.12[34589]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1734194055.7196329-15462-153591703695286/.source.kube dest=/etc/containers/systemd/quadlet-demo.kube owner=root group=0 mode=0644 _original_basename=quadlet-demo.kube follow=False checksum=7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:34:16 managed-node3 python3.12[34720]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:34:16 managed-node3 systemd[1]: Reload requested from client PID 34721 ('systemctl') (unit session-5.scope)... Dec 14 11:34:16 managed-node3 systemd[1]: Reloading... Dec 14 11:34:17 managed-node3 systemd[1]: Reloading finished in 210 ms. Dec 14 11:34:17 managed-node3 python3.12[34904]: ansible-systemd Invoked with name=quadlet-demo.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None Dec 14 11:34:17 managed-node3 systemd[1]: Starting quadlet-demo.service... ░░ Subject: A start job for unit quadlet-demo.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo.service has begun execution. ░░ ░░ The job identifier is 2452. Dec 14 11:34:17 managed-node3 quadlet-demo[34908]: time="2024-12-14T11:34:17-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" Dec 14 11:34:17 managed-node3 quadlet-demo[34908]: Pods stopped: Dec 14 11:34:17 managed-node3 quadlet-demo[34908]: Pods removed: Dec 14 11:34:17 managed-node3 quadlet-demo[34908]: Secrets removed: Dec 14 11:34:17 managed-node3 quadlet-demo[34908]: Volumes removed: Dec 14 11:34:17 managed-node3 podman[34908]: 2024-12-14 11:34:17.734699606 -0500 EST m=+0.029844560 volume create wp-pv-claim Dec 14 11:34:17 managed-node3 quadlet-demo[34908]: Error: image for service container: building local pause image: finding pause binary: exec: "catatonit": executable file not found in $PATH Dec 14 11:34:17 managed-node3 systemd[1]: quadlet-demo.service: Main process exited, code=exited, status=125/n/a ░░ Subject: Unit process exited ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ An ExecStart= process belonging to unit quadlet-demo.service has exited. ░░ ░░ The process' exit code is 'exited' and its exit status is 125. Dec 14 11:34:17 managed-node3 quadlet-demo[34916]: time="2024-12-14T11:34:17-05:00" level=warning msg="Failed to decode the keys [\"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options\" \"storage.options.overlay.pull_options.enable_partial_images\" \"storage.options.overlay.pull_options.use_hard_links\" \"storage.options.overlay.pull_options.ostree_repos\" \"storage.options.overlay.pull_options.convert_images\"] from \"/usr/share/containers/storage.conf\"" Dec 14 11:34:17 managed-node3 quadlet-demo[34916]: Pods stopped: Dec 14 11:34:17 managed-node3 quadlet-demo[34916]: Pods removed: Dec 14 11:34:17 managed-node3 quadlet-demo[34916]: Secrets removed: Dec 14 11:34:17 managed-node3 quadlet-demo[34916]: Volumes removed: Dec 14 11:34:17 managed-node3 systemd[1]: quadlet-demo.service: Failed with result 'exit-code'. ░░ Subject: Unit failed ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit quadlet-demo.service has entered the 'failed' state with result 'exit-code'. Dec 14 11:34:17 managed-node3 systemd[1]: Failed to start quadlet-demo.service. ░░ Subject: A start job for unit quadlet-demo.service has failed ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit quadlet-demo.service has finished with a failure. ░░ ░░ The job identifier is 2452 and the job result is failed. Dec 14 11:34:18 managed-node3 python3.12[35056]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:34:18 managed-node3 python3.12[35188]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps -a _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:34:19 managed-node3 python3.12[35326]: ansible-ansible.legacy.command Invoked with _raw_params=podman pod ps --ctr-ids --ctr-names --ctr-status _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:34:19 managed-node3 python3.12[35465]: ansible-ansible.legacy.command Invoked with _raw_params=set -euo pipefail; systemctl list-units --all | grep quadlet _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:34:19 managed-node3 python3.12[35599]: ansible-ansible.legacy.command Invoked with _raw_params=ls -alrtF /etc/systemd/system _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:34:21 managed-node3 python3.12[35862]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:34:22 managed-node3 python3.12[35999]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:34:25 managed-node3 python3.12[36132]: ansible-ansible.legacy.dnf Invoked with name=['firewalld'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Dec 14 11:34:26 managed-node3 python3.12[36264]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False scope=system no_block=False state=None enabled=None force=None Dec 14 11:34:26 managed-node3 python3.12[36397]: ansible-ansible.legacy.systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False scope=system no_block=False force=None masked=None Dec 14 11:34:27 managed-node3 python3.12[36530]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Dec 14 11:34:27 managed-node3 python3.12[36661]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Dec 14 11:34:31 managed-node3 python3.12[37208]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:34:32 managed-node3 python3.12[37341]: ansible-systemd Invoked with name=quadlet-demo.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None Dec 14 11:34:32 managed-node3 systemd[1]: Reload requested from client PID 37344 ('systemctl') (unit session-5.scope)... Dec 14 11:34:32 managed-node3 systemd[1]: Reloading... Dec 14 11:34:33 managed-node3 systemd[1]: Reloading finished in 212 ms. Dec 14 11:34:33 managed-node3 python3.12[37527]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.kube follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:34:34 managed-node3 python3.12[37791]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.kube state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:34:35 managed-node3 python3.12[37922]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:34:35 managed-node3 systemd[1]: Reload requested from client PID 37923 ('systemctl') (unit session-5.scope)... Dec 14 11:34:35 managed-node3 systemd[1]: Reloading... Dec 14 11:34:35 managed-node3 systemd[1]: Reloading finished in 206 ms. Dec 14 11:34:35 managed-node3 python3.12[38107]: ansible-ansible.legacy.command Invoked with _raw_params=podman image prune --all -f _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:34:36 managed-node3 podman[38108]: 2024-12-14 11:34:36.318790679 -0500 EST m=+0.571931776 image untag fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b quay.io/linux-system-roles/wordpress:4.8-apache Dec 14 11:34:36 managed-node3 podman[38108]: 2024-12-14 11:34:35.768535033 -0500 EST m=+0.021676199 image remove fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b Dec 14 11:34:36 managed-node3 podman[38108]: 2024-12-14 11:34:36.449501256 -0500 EST m=+0.702642349 image untag 5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d quay.io/linux-system-roles/envoyproxy:v1.25.0 Dec 14 11:34:36 managed-node3 podman[38108]: 2024-12-14 11:34:36.318804387 -0500 EST m=+0.571945504 image remove 5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d Dec 14 11:34:37 managed-node3 python3.12[38246]: ansible-ansible.legacy.command Invoked with _raw_params=podman images -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:34:37 managed-node3 python3.12[38384]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume ls -n _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:34:37 managed-node3 python3.12[38522]: ansible-ansible.legacy.command Invoked with _raw_params=podman ps --noheading _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:34:38 managed-node3 python3.12[38661]: ansible-ansible.legacy.command Invoked with _raw_params=podman network ls -n -q _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:34:39 managed-node3 python3.12[39076]: ansible-service_facts Invoked Dec 14 11:34:42 managed-node3 python3.12[39314]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:34:43 managed-node3 python3.12[39447]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.yml follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Dec 14 11:34:44 managed-node3 python3.12[39711]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Dec 14 11:34:45 managed-node3 python3.12[39842]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None Dec 14 11:34:45 managed-node3 systemd[1]: Reload requested from client PID 39843 ('systemctl') (unit session-5.scope)... Dec 14 11:34:45 managed-node3 systemd[1]: Reloading... Dec 14 11:34:45 managed-node3 systemd[1]: Reloading finished in 206 ms. Dec 14 11:34:46 managed-node3 podman[40143]: 2024-12-14 11:34:46.321105394 -0500 EST m=+0.080083267 container health_status e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72 (image=quay.io/linux-system-roles/mysql:5.6, name=quadlet-demo-mysql, health_status=healthy, health_failing_streak=1, health_log=, PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service) Dec 14 11:34:46 managed-node3 systemd[1]: e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72-7b63df59ce7fdf47.service: Main process exited, code=exited, status=125/n/a ░░ Subject: Unit process exited ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ An ExecStart= process belonging to unit e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72-7b63df59ce7fdf47.service has exited. ░░ ░░ The process' exit code is 'exited' and its exit status is 125. Dec 14 11:34:46 managed-node3 systemd[1]: e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72-7b63df59ce7fdf47.service: Failed with result 'exit-code'. ░░ Subject: Unit failed ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit e8a2f6c7683ed26ffd7bb194b722d1f17be14f9011be2b344ddafa6eafbbaf72-7b63df59ce7fdf47.service has entered the 'failed' state with result 'exit-code'. Dec 14 11:34:46 managed-node3 podman[40315]: 2024-12-14 11:34:46.853358431 -0500 EST m=+0.022928787 volume remove wp-pv-claim Dec 14 11:34:47 managed-node3 python3.12[40452]: ansible-ansible.legacy.command Invoked with _raw_params=exec 1>&2 set -x set -o pipefail systemctl list-units --plain -l --all | grep quadlet || : systemctl list-unit-files --all | grep quadlet || : systemctl list-units --plain --failed -l --all | grep quadlet || : _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Dec 14 11:34:48 managed-node3 python3.12[40590]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None PLAY RECAP ********************************************************************* managed-node3 : ok=262 changed=32 unreachable=0 failed=2 skipped=287 rescued=2 ignored=0 TASKS RECAP ******************************************************************** Saturday 14 December 2024 11:34:48 -0500 (0:00:00.416) 0:01:57.225 ***** =============================================================================== fedora.linux_system_roles.podman : Ensure container images are present -- 18.12s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 fedora.linux_system_roles.podman : Ensure container images are present --- 7.25s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 fedora.linux_system_roles.certificate : Ensure provider packages are installed --- 3.06s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed --- 2.63s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 fedora.linux_system_roles.podman : For testing and debugging - services --- 2.19s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:197 Gathering Facts --------------------------------------------------------- 1.48s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:9 fedora.linux_system_roles.certificate : Slurp the contents of the files --- 1.38s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152 fedora.linux_system_roles.podman : Gather the package facts ------------- 1.37s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.podman : Remove volumes ----------------------- 1.33s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:99 fedora.linux_system_roles.certificate : Remove files -------------------- 1.30s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181 fedora.linux_system_roles.firewall : Configure firewall ----------------- 1.30s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 fedora.linux_system_roles.certificate : Ensure provider service is running --- 1.27s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 fedora.linux_system_roles.podman : Prune images no longer in use -------- 1.16s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:120 fedora.linux_system_roles.firewall : Enable and start firewalld service --- 1.10s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 fedora.linux_system_roles.podman : Ensure quadlet file content is present --- 1.04s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 fedora.linux_system_roles.firewall : Configure firewall ----------------- 1.01s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 fedora.linux_system_roles.podman : Start service ------------------------ 0.95s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 fedora.linux_system_roles.certificate : Ensure certificate requests ----- 0.91s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101 fedora.linux_system_roles.firewall : Install firewalld ------------------ 0.89s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 fedora.linux_system_roles.podman : Ensure quadlet file is copied -------- 0.86s /tmp/collections-9ny/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48