ansible-playbook 2.9.27 config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.9/site-packages/ansible executable location = /usr/local/bin/ansible-playbook python version = 3.9.19 (main, May 16 2024, 11:40:09) [GCC 8.5.0 20210514 (Red Hat 8.5.0-22)] No config file found; using defaults [WARNING]: running playbook inside collection fedora.linux_system_roles Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_quadlet_demo.yml *********************************************** 2 plays in /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml PLAY [all] ********************************************************************* META: ran handlers TASK [Include vault variables] ************************************************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:5 Wednesday 31 July 2024 21:23:46 -0400 (0:00:00.020) 0:00:00.020 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_test_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n35383939616163653333633431363463313831383037386236646138333162396161356130303461\n3932623930643263313563336163316337643562333936360a363538636631313039343233383732\n38666530383538656639363465313230343533386130303833336434303438333161656262346562\n3362626538613031640a663330613638366132356534363534353239616666653466353961323533\n6565\n" }, "mysql_container_root_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n61333932373230333539663035366431326163363166363036323963623131363530326231303634\n6635326161643165363366323062333334363730376631660a393566366139353861656364656661\n38653463363837336639363032646433666361646535366137303464623261313663643336306465\n6264663730656337310a343962353137386238383064646533366433333437303566656433386233\n34343235326665646661623131643335313236313131353661386338343366316261643634653633\n3832313034366536616531323963333234326461353130303532\n" } }, "ansible_included_var_files": [ "/var/ARTIFACTS/work-generalxrr7nq91/plans/general/tree/podman/tests/vars/vault-variables.yml" ], "changed": false } META: ran handlers META: ran handlers PLAY [Deploy the quadlet demo app] ********************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:9 Wednesday 31 July 2024 21:23:46 -0400 (0:00:00.016) 0:00:00.037 ******** ok: [managed_node1] META: ran handlers TASK [Generate certificates] *************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:39 Wednesday 31 July 2024 21:23:47 -0400 (0:00:00.891) 0:00:00.928 ******** TASK [fedora.linux_system_roles.certificate : Set version specific variables] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:2 Wednesday 31 July 2024 21:23:47 -0400 (0:00:00.032) 0:00:00.961 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml for managed_node1 TASK [fedora.linux_system_roles.certificate : Ensure ansible_facts used by role] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:2 Wednesday 31 July 2024 21:23:47 -0400 (0:00:00.018) 0:00:00.980 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.certificate : Check if system is ostree] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:10 Wednesday 31 July 2024 21:23:47 -0400 (0:00:00.035) 0:00:01.015 ******** ok: [managed_node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.certificate : Set flag to indicate system is ostree] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:15 Wednesday 31 July 2024 21:23:48 -0400 (0:00:00.458) 0:00:01.474 ******** ok: [managed_node1] => { "ansible_facts": { "__certificate_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.certificate : Set platform/version specific variables] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/set_vars.yml:19 Wednesday 31 July 2024 21:23:48 -0400 (0:00:00.035) 0:00:01.509 ******** skipping: [managed_node1] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=CentOS_8.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=CentOS_8.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 Wednesday 31 July 2024 21:23:48 -0400 (0:00:00.049) 0:00:01.559 ******** ok: [managed_node1] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.certificate : Ensure provider packages are installed] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 Wednesday 31 July 2024 21:23:51 -0400 (0:00:02.912) 0:00:04.472 ******** ok: [managed_node1] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.certificate : Ensure pre-scripts hooks directory exists] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:35 Wednesday 31 July 2024 21:23:53 -0400 (0:00:02.440) 0:00:06.912 ******** ok: [managed_node1] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": false, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//pre-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure post-scripts hooks directory exists] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:61 Wednesday 31 July 2024 21:23:53 -0400 (0:00:00.466) 0:00:07.379 ******** ok: [managed_node1] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": false, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/etc/certmonger//post-scripts", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.certificate : Ensure provider service is running] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:90 Wednesday 31 July 2024 21:23:54 -0400 (0:00:00.369) 0:00:07.748 ******** ok: [managed_node1] => (item=certmonger) => { "__certificate_provider": "certmonger", "ansible_loop_var": "__certificate_provider", "changed": false, "enabled": true, "name": "certmonger", "state": "started", "status": { "ActiveEnterTimestamp": "Wed 2024-07-31 21:19:45 EDT", "ActiveEnterTimestampMonotonic": "8181455", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "dbus.socket network.target systemd-journald.socket basic.target dbus.service syslog.target system.slice sysinit.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-31 21:19:45 EDT", "AssertTimestampMonotonic": "8095896", "Before": "multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedorahosted.certmonger", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-31 21:19:45 EDT", "ConditionTimestampMonotonic": "8095895", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroup": "/system.slice/certmonger.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Certificate monitoring and PKI enrollment", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "EnvironmentFiles": "/etc/sysconfig/certmonger (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "675", "ExecMainStartTimestamp": "Wed 2024-07-31 21:19:45 EDT", "ExecMainStartTimestampMonotonic": "8097978", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/certmonger ; argv[]=/usr/sbin/certmonger -S -p /run/certmonger.pid -n $OPTS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/certmonger.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "certmonger.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-31 21:19:45 EDT", "InactiveExitTimestampMonotonic": "8098017", "InvocationID": "4c7e9e9f61b94b478682b704eada7207", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "675", "MemoryAccounting": "yes", "MemoryCurrent": "8548352", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "certmonger.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PIDFile": "/run/certmonger.pid", "PartOf": "dbus.service", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "dbus.socket system.slice sysinit.target", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Wed 2024-07-31 21:19:45 EDT", "StateChangeTimestampMonotonic": "8181455", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "1", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogTimestamp": "Wed 2024-07-31 21:19:45 EDT", "WatchdogTimestampMonotonic": "8181453", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.certificate : Ensure certificate requests] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:101 Wednesday 31 July 2024 21:23:54 -0400 (0:00:00.678) 0:00:08.427 ******** changed: [managed_node1] => (item={'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}) => { "ansible_loop_var": "item", "changed": true, "item": { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } } MSG: Certificate requested (new). TASK [fedora.linux_system_roles.certificate : Slurp the contents of the files] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152 Wednesday 31 July 2024 21:23:55 -0400 (0:00:00.914) 0:00:09.341 ******** ok: [managed_node1] => (item=['cert', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnekNDQW11Z0F3SUJBZ0lSQVBicEV0SWx5RVZjallVMjhOTFN0NjB3RFFZSktvWklodmNOQVFFTEJRQXcKVURFZ01CNEdBMVVFQXd3WFRHOWpZV3dnVTJsbmJtbHVaeUJCZFhSb2IzSnBkSGt4TERBcUJnTlZCQU1NSTJZMgpaVGt4TW1ReUxUSTFZemcwTlRWakxUaGtPRFV6Tm1Zd0xXUXlaREppTjJGaU1CNFhEVEkwTURnd01UQXhNak0xCk5Wb1hEVEkxTURnd01UQXhNRGt4TWxvd0ZERVNNQkFHQTFVRUF4TUpiRzlqWVd4b2IzTjBNSUlCSWpBTkJna3EKaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUExNzVoODVxSSswSXJoTVprQlplYnE3UXJMTXl6QmdRZwpMbVRta2RYa3BBQlpHL0xORWpPdW5BeVlBQmxDaGsxeE5WdlcwSEpqRlJ4enJ5V0lDZGQyM0dxNER4Z21qR2tEClpSSVNGaDFDVm5XTlUwZDF4NFQ4bFliTGVSNmF4YWxmK1dsV01BSkR3NjhmdFN2QzBldlY4TWNReXZCMlF5K2MKTzBMZVBLVWd6NTNrMC8xb2tmYTRvNGlNQzl2UmV3b0xjZk9ZTHFYV0NiSlgwTCsxZ3hoK1dWbVdoZHhHMUlsQQpYWlNmQ05yVlRxaEVMQ3YxeGNRRk5FL3V2VEt2bm90Rytyd1hhcndsUUptc1RtMmozRmR5TDlXNnE0eW5ra2tjCnErUTRwS1g2TXNRSlV6M093OUJZQkpkd1pWb3p4Uzltd0xCazJVbXp4ZDdvWFNhcHgwM0NDUUlEQVFBQm80R1QKTUlHUU1Bc0dBMVVkRHdRRUF3SUZvREFVQmdOVkhSRUVEVEFMZ2dsc2IyTmhiR2h2YzNRd0hRWURWUjBsQkJZdwpGQVlJS3dZQkJRVUhBd0VHQ0NzR0FRVUZCd01DTUF3R0ExVWRFd0VCL3dRQ01BQXdIUVlEVlIwT0JCWUVGRDIwCnpOMGJ3RHptb1JoY3ZNQldRTVpTVlg2cU1COEdBMVVkSXdRWU1CYUFGTCtkV1RqNXBBMVA3aXp0WENsb2x0VWYKYXRCNk1BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQWhzeHdrODhKLzlUS2FleTVKTk5GM1A2bzM3Um9kTU5BNAorOXNHS2cwQ2Q4WEJGZURScTE5Y0RjMEZ1TWdyRXI2TnJuWDc4MnRsaE1WNXNhZWFJMUNTWm9jb0lUWUpVbDF6ClN3MDltMjhwaWd0SXIwc1VadXRiWi9MdjRTRGxwWWttTlJGeXlJTXRTdmVtVHQ0SlpkRzF3dDVpOElrZnVoaVcKczRnS1JjQ1VsSWVVRzd2eHNKNUlnNnNiT0ppenRzOE1KMi9TcDRRdGZJRzVpNEx0aGpOT1FnQnBObWQ4STh3agpNOTQzMm5Ha3R5eGxzeE9tbFR5RUFXS1BMTjI5VWFiL2FyUHh6QUNYa3d5eW1xTExGdXIrUk0rZFFSTncvUHVYCldGd2xvK1ZBRUg5dlFzbVFZRTFYbHJ4UVBBUlNDNlMzeGI2Mk1PZHVJRUV6UlhLQU5FZDkKLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "cert", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } ok: [managed_node1] => (item=['key', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBQUklWQVRFIEtFWS0tLS0tCk1JSUV2Z0lCQURBTkJna3Foa2lHOXcwQkFRRUZBQVNDQktnd2dnU2tBZ0VBQW9JQkFRRFh2bUh6bW9qN1FpdUUKeG1RRmw1dXJ0Q3NzekxNR0JDQXVaT2FSMWVTa0FGa2I4czBTTTY2Y0RKZ0FHVUtHVFhFMVc5YlFjbU1WSEhPdgpKWWdKMTNiY2FyZ1BHQ2FNYVFObEVoSVdIVUpXZFkxVFIzWEhoUHlWaHN0NUhwckZxVi81YVZZd0FrUERyeCsxCks4TFI2OVh3eHhESzhIWkRMNXc3UXQ0OHBTRFBuZVRUL1dpUjlyaWppSXdMMjlGN0NndHg4NWd1cGRZSnNsZlEKdjdXREdINVpXWmFGM0ViVWlVQmRsSjhJMnRWT3FFUXNLL1hGeEFVMFQrNjlNcStlaTBiNnZCZHF2Q1ZBbWF4TwpiYVBjVjNJdjFicXJqS2VTU1J5cjVEaWtwZm95eEFsVFBjN0QwRmdFbDNCbFdqUEZMMmJBc0dUWlNiUEYzdWhkCkpxbkhUY0lKQWdNQkFBRUNnZ0VCQUtmTm5xWTVCbDVYZ3ozTWlUT0lUajVvN2tQMEd6S010dk1lNis5MmJGWGIKSjNRRGo5bWVkU3RPNkZMN1VyRGZQelIvY0QvRkdLZnM3MVhGWFdBNlJHc0FYcFR1OWR5ME9UMjlwbStvNWxZVgpMSTBmd2M0ajg4bmhoQlY2NW1yamFKQzNpNS9md3hGSy91Q05Yd1NqNkUxVXFBMFFwd3R1MXBlSlZRM2FhZ0RRCm9xS3dqL1M5NGkrVU1yOTRxWG1XZWtpbGloYlFDbFZSVFB4d0pVbzl5QUFHUDRJbFA2enJ4ejk4Q0dsT08rZUQKaUVNOXlJd3J3Z05VMWNIN1YzbkNRMUo1S3hjaWtYazgwUWZKRXNEU1o0eFA0dHQ1VSs0dFBRMExyMTdSeXNmbgp1VzZlVGNnSG1TY2FrQ3lXM3FVVm1scTZRZEpFSWcyenROL2tTMjhZL2xrQ2dZRUE5WGNBMUhrdlV5TkVPNnBECjdEZmtXbEp4ZWVJekpvdlhZd005TnpkZzRpOXZ5ajkrKzMvM01wZEV2RCt1RDU0NTFsMHM3LzFSQit5Uk1uaDAKY2t5TnkvdFJoS0xudkNHZVpZVGlEdnVrOEhyNkpWSGJrY3orcDI4dmozNU15RHpWczNLYUlpaHRtOUEwdmFYKwo5azE2MVRONFRtcWdWZzZhd1pENERpbGVFSDhDZ1lFQTRRRFM1cndyd1QwWk03bjAwaTFDeCtzSENwbVQ2cDJsCnYvaUxSZ0ZmcTY3bUozbTNva2lIdFJFWWd4RVBHRjJKelFabWg5RTAwR0w3SllvRGJYb3k3bFZJYWk0SzZIdnoKVWU3NFRIcUtZMElzb3FScEh0QWtsYXM2MmRvdEp4cXcwOGdTZmN4M0pqMDUyZjYxK1hVbklJVmtPRDBuUFhiZgppTmVsbGlBd2FYY0NnWUFCbnl5cE9yZStVK0JTaUpoMFNoWS9Kb1pMYU53dnFBTGZTNWt1OFRXNXFOaVRQSUlwClJUb3I0bVg0M0c3WStaR21RZTFYRkpSaGVnV25rTFlzUHljOHJzZG1HZkZ3QThUWkpOeWRmcm9qUUp6QWZGNzQKQXFPNEZYcjU5cnFwSkFYQmJzTm44RjV4QmhvakdxS3BtKzl3MHlzN1VXY0h5b2JwT256NUV2cmxId0tCZ1FDcgptSkkvY0JDNmpkc3pqTVUrcVUvOTRhdDhGUVl1N3MzaTZ5MWtWVlhlUEVQQXBWNDhBSzZtazRRcktRNW5HYWNuCldwTkZUM0xFM2tnRktBMVhKOVBCMmY2aEk3NGo4YXZ1dDU1YnpUVFBZRDlrUnBqNDYrcUt6b3pTYWtiM1JLN0kKeFJQOUpydFR5V3ZlYWhma1NNbHZUdXZRekNxcUFJZVgrRWFMaUw4REh3S0JnQmQ5ZFI3WkF6ek9sSE55djJ0cQpmSlVQN3ZrYTJ0VVhTdlM5aHYwb3dJQnlmV2FQZzZFSUVDY2lnam42bGpueElNRkpGNFVRdXE1c0lRSXBwd0l4CjZUTEx5SnFucThmWDRjWGh4dU13VHZPWm5GMWo2cnN1L1J4ZmhzZ0sweFVwUGtuQUVPbnZkRlIvV0JkZzlPOS8KTVBoM2ZSVkl0czgzdUxTY0tNTU1rNzlTCi0tLS0tRU5EIFBSSVZBVEUgS0VZLS0tLS0K", "encoding": "base64", "item": [ "key", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/private/quadlet_demo.key" } ok: [managed_node1] => (item=['ca', {'name': 'quadlet_demo', 'dns': ['localhost'], 'ca': 'self-sign'}]) => { "ansible_loop_var": "item", "changed": false, "content": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURnekNDQW11Z0F3SUJBZ0lSQVBicEV0SWx5RVZjallVMjhOTFN0NjB3RFFZSktvWklodmNOQVFFTEJRQXcKVURFZ01CNEdBMVVFQXd3WFRHOWpZV3dnVTJsbmJtbHVaeUJCZFhSb2IzSnBkSGt4TERBcUJnTlZCQU1NSTJZMgpaVGt4TW1ReUxUSTFZemcwTlRWakxUaGtPRFV6Tm1Zd0xXUXlaREppTjJGaU1CNFhEVEkwTURnd01UQXhNak0xCk5Wb1hEVEkxTURnd01UQXhNRGt4TWxvd0ZERVNNQkFHQTFVRUF4TUpiRzlqWVd4b2IzTjBNSUlCSWpBTkJna3EKaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUExNzVoODVxSSswSXJoTVprQlplYnE3UXJMTXl6QmdRZwpMbVRta2RYa3BBQlpHL0xORWpPdW5BeVlBQmxDaGsxeE5WdlcwSEpqRlJ4enJ5V0lDZGQyM0dxNER4Z21qR2tEClpSSVNGaDFDVm5XTlUwZDF4NFQ4bFliTGVSNmF4YWxmK1dsV01BSkR3NjhmdFN2QzBldlY4TWNReXZCMlF5K2MKTzBMZVBLVWd6NTNrMC8xb2tmYTRvNGlNQzl2UmV3b0xjZk9ZTHFYV0NiSlgwTCsxZ3hoK1dWbVdoZHhHMUlsQQpYWlNmQ05yVlRxaEVMQ3YxeGNRRk5FL3V2VEt2bm90Rytyd1hhcndsUUptc1RtMmozRmR5TDlXNnE0eW5ra2tjCnErUTRwS1g2TXNRSlV6M093OUJZQkpkd1pWb3p4Uzltd0xCazJVbXp4ZDdvWFNhcHgwM0NDUUlEQVFBQm80R1QKTUlHUU1Bc0dBMVVkRHdRRUF3SUZvREFVQmdOVkhSRUVEVEFMZ2dsc2IyTmhiR2h2YzNRd0hRWURWUjBsQkJZdwpGQVlJS3dZQkJRVUhBd0VHQ0NzR0FRVUZCd01DTUF3R0ExVWRFd0VCL3dRQ01BQXdIUVlEVlIwT0JCWUVGRDIwCnpOMGJ3RHptb1JoY3ZNQldRTVpTVlg2cU1COEdBMVVkSXdRWU1CYUFGTCtkV1RqNXBBMVA3aXp0WENsb2x0VWYKYXRCNk1BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQWhzeHdrODhKLzlUS2FleTVKTk5GM1A2bzM3Um9kTU5BNAorOXNHS2cwQ2Q4WEJGZURScTE5Y0RjMEZ1TWdyRXI2TnJuWDc4MnRsaE1WNXNhZWFJMUNTWm9jb0lUWUpVbDF6ClN3MDltMjhwaWd0SXIwc1VadXRiWi9MdjRTRGxwWWttTlJGeXlJTXRTdmVtVHQ0SlpkRzF3dDVpOElrZnVoaVcKczRnS1JjQ1VsSWVVRzd2eHNKNUlnNnNiT0ppenRzOE1KMi9TcDRRdGZJRzVpNEx0aGpOT1FnQnBObWQ4STh3agpNOTQzMm5Ha3R5eGxzeE9tbFR5RUFXS1BMTjI5VWFiL2FyUHh6QUNYa3d5eW1xTExGdXIrUk0rZFFSTncvUHVYCldGd2xvK1ZBRUg5dlFzbVFZRTFYbHJ4UVBBUlNDNlMzeGI2Mk1PZHVJRUV6UlhLQU5FZDkKLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=", "encoding": "base64", "item": [ "ca", { "ca": "self-sign", "dns": [ "localhost" ], "name": "quadlet_demo" } ], "source": "/etc/pki/tls/certs/quadlet_demo.crt" } TASK [fedora.linux_system_roles.certificate : Create return data] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:160 Wednesday 31 July 2024 21:23:56 -0400 (0:00:01.083) 0:00:10.424 ******** ok: [managed_node1] => { "ansible_facts": { "certificate_test_certs": { "quadlet_demo": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRAPbpEtIlyEVcjYU28NLSt60wDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Y2\nZTkxMmQyLTI1Yzg0NTVjLThkODUzNmYwLWQyZDJiN2FiMB4XDTI0MDgwMTAxMjM1\nNVoXDTI1MDgwMTAxMDkxMlowFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA175h85qI+0IrhMZkBZebq7QrLMyzBgQg\nLmTmkdXkpABZG/LNEjOunAyYABlChk1xNVvW0HJjFRxzryWICdd23Gq4DxgmjGkD\nZRISFh1CVnWNU0d1x4T8lYbLeR6axalf+WlWMAJDw68ftSvC0evV8McQyvB2Qy+c\nO0LePKUgz53k0/1okfa4o4iMC9vRewoLcfOYLqXWCbJX0L+1gxh+WVmWhdxG1IlA\nXZSfCNrVTqhELCv1xcQFNE/uvTKvnotG+rwXarwlQJmsTm2j3FdyL9W6q4ynkkkc\nq+Q4pKX6MsQJUz3Ow9BYBJdwZVozxS9mwLBk2Umzxd7oXSapx03CCQIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFD20\nzN0bwDzmoRhcvMBWQMZSVX6qMB8GA1UdIwQYMBaAFL+dWTj5pA1P7iztXCloltUf\natB6MA0GCSqGSIb3DQEBCwUAA4IBAQAhsxwk88J/9TKaey5JNNF3P6o37RodMNA4\n+9sGKg0Cd8XBFeDRq19cDc0FuMgrEr6NrnX782tlhMV5saeaI1CSZocoITYJUl1z\nSw09m28pigtIr0sUZutbZ/Lv4SDlpYkmNRFyyIMtSvemTt4JZdG1wt5i8IkfuhiW\ns4gKRcCUlIeUG7vxsJ5Ig6sbOJizts8MJ2/Sp4QtfIG5i4LthjNOQgBpNmd8I8wj\nM9432nGktyxlsxOmlTyEAWKPLN29Uab/arPxzACXkwyymqLLFur+RM+dQRNw/PuX\nWFwlo+VAEH9vQsmQYE1XlrxQPARSC6S3xb62MOduIEEzRXKANEd9\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRAPbpEtIlyEVcjYU28NLSt60wDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Y2\nZTkxMmQyLTI1Yzg0NTVjLThkODUzNmYwLWQyZDJiN2FiMB4XDTI0MDgwMTAxMjM1\nNVoXDTI1MDgwMTAxMDkxMlowFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA175h85qI+0IrhMZkBZebq7QrLMyzBgQg\nLmTmkdXkpABZG/LNEjOunAyYABlChk1xNVvW0HJjFRxzryWICdd23Gq4DxgmjGkD\nZRISFh1CVnWNU0d1x4T8lYbLeR6axalf+WlWMAJDw68ftSvC0evV8McQyvB2Qy+c\nO0LePKUgz53k0/1okfa4o4iMC9vRewoLcfOYLqXWCbJX0L+1gxh+WVmWhdxG1IlA\nXZSfCNrVTqhELCv1xcQFNE/uvTKvnotG+rwXarwlQJmsTm2j3FdyL9W6q4ynkkkc\nq+Q4pKX6MsQJUz3Ow9BYBJdwZVozxS9mwLBk2Umzxd7oXSapx03CCQIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFD20\nzN0bwDzmoRhcvMBWQMZSVX6qMB8GA1UdIwQYMBaAFL+dWTj5pA1P7iztXCloltUf\natB6MA0GCSqGSIb3DQEBCwUAA4IBAQAhsxwk88J/9TKaey5JNNF3P6o37RodMNA4\n+9sGKg0Cd8XBFeDRq19cDc0FuMgrEr6NrnX782tlhMV5saeaI1CSZocoITYJUl1z\nSw09m28pigtIr0sUZutbZ/Lv4SDlpYkmNRFyyIMtSvemTt4JZdG1wt5i8IkfuhiW\ns4gKRcCUlIeUG7vxsJ5Ig6sbOJizts8MJ2/Sp4QtfIG5i4LthjNOQgBpNmd8I8wj\nM9432nGktyxlsxOmlTyEAWKPLN29Uab/arPxzACXkwyymqLLFur+RM+dQRNw/PuX\nWFwlo+VAEH9vQsmQYE1XlrxQPARSC6S3xb62MOduIEEzRXKANEd9\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvgIBADANBgkqhkiG9w0BAQEFAASCBKgwggSkAgEAAoIBAQDXvmHzmoj7QiuE\nxmQFl5urtCsszLMGBCAuZOaR1eSkAFkb8s0SM66cDJgAGUKGTXE1W9bQcmMVHHOv\nJYgJ13bcargPGCaMaQNlEhIWHUJWdY1TR3XHhPyVhst5HprFqV/5aVYwAkPDrx+1\nK8LR69XwxxDK8HZDL5w7Qt48pSDPneTT/WiR9rijiIwL29F7Cgtx85gupdYJslfQ\nv7WDGH5ZWZaF3EbUiUBdlJ8I2tVOqEQsK/XFxAU0T+69Mq+ei0b6vBdqvCVAmaxO\nbaPcV3Iv1bqrjKeSSRyr5DikpfoyxAlTPc7D0FgEl3BlWjPFL2bAsGTZSbPF3uhd\nJqnHTcIJAgMBAAECggEBAKfNnqY5Bl5Xgz3MiTOITj5o7kP0GzKMtvMe6+92bFXb\nJ3QDj9medStO6FL7UrDfPzR/cD/FGKfs71XFXWA6RGsAXpTu9dy0OT29pm+o5lYV\nLI0fwc4j88nhhBV65mrjaJC3i5/fwxFK/uCNXwSj6E1UqA0Qpwtu1peJVQ3aagDQ\noqKwj/S94i+UMr94qXmWekilihbQClVRTPxwJUo9yAAGP4IlP6zrxz98CGlOO+eD\niEM9yIwrwgNU1cH7V3nCQ1J5KxcikXk80QfJEsDSZ4xP4tt5U+4tPQ0Lr17Rysfn\nuW6eTcgHmScakCyW3qUVmlq6QdJEIg2ztN/kS28Y/lkCgYEA9XcA1HkvUyNEO6pD\n7DfkWlJxeeIzJovXYwM9Nzdg4i9vyj9++3/3MpdEvD+uD5451l0s7/1RB+yRMnh0\nckyNy/tRhKLnvCGeZYTiDvuk8Hr6JVHbkcz+p28vj35MyDzVs3KaIihtm9A0vaX+\n9k161TN4TmqgVg6awZD4DileEH8CgYEA4QDS5rwrwT0ZM7n00i1Cx+sHCpmT6p2l\nv/iLRgFfq67mJ3m3okiHtREYgxEPGF2JzQZmh9E00GL7JYoDbXoy7lVIai4K6Hvz\nUe74THqKY0IsoqRpHtAklas62dotJxqw08gSfcx3Jj052f61+XUnIIVkOD0nPXbf\niNelliAwaXcCgYABnyypOre+U+BSiJh0ShY/JoZLaNwvqALfS5ku8TW5qNiTPIIp\nRTor4mX43G7Y+ZGmQe1XFJRhegWnkLYsPyc8rsdmGfFwA8TZJNydfrojQJzAfF74\nAqO4FXr59rqpJAXBbsNn8F5xBhojGqKpm+9w0ys7UWcHyobpOnz5EvrlHwKBgQCr\nmJI/cBC6jdszjMU+qU/94at8FQYu7s3i6y1kVVXePEPApV48AK6mk4QrKQ5nGacn\nWpNFT3LE3kgFKA1XJ9PB2f6hI74j8avut55bzTTPYD9kRpj46+qKzozSakb3RK7I\nxRP9JrtTyWveahfkSMlvTuvQzCqqAIeX+EaLiL8DHwKBgBd9dR7ZAzzOlHNyv2tq\nfJUP7vka2tUXSvS9hv0owIByfWaPg6EIECcigjn6ljnxIMFJF4UQuq5sIQIppwIx\n6TLLyJqnq8fX4cXhxuMwTvOZnF1j6rsu/RxfhsgK0xUpPknAEOnvdFR/WBdg9O9/\nMPh3fRVIts83uLScKMMMk79S\n-----END PRIVATE KEY-----\n" } } }, "changed": false } TASK [fedora.linux_system_roles.certificate : Stop tracking certificates] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:176 Wednesday 31 July 2024 21:23:57 -0400 (0:00:00.039) 0:00:10.464 ******** ok: [managed_node1] => (item={'cert': '/etc/pki/tls/certs/quadlet_demo.crt', 'cert_content': '-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRAPbpEtIlyEVcjYU28NLSt60wDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Y2\nZTkxMmQyLTI1Yzg0NTVjLThkODUzNmYwLWQyZDJiN2FiMB4XDTI0MDgwMTAxMjM1\nNVoXDTI1MDgwMTAxMDkxMlowFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA175h85qI+0IrhMZkBZebq7QrLMyzBgQg\nLmTmkdXkpABZG/LNEjOunAyYABlChk1xNVvW0HJjFRxzryWICdd23Gq4DxgmjGkD\nZRISFh1CVnWNU0d1x4T8lYbLeR6axalf+WlWMAJDw68ftSvC0evV8McQyvB2Qy+c\nO0LePKUgz53k0/1okfa4o4iMC9vRewoLcfOYLqXWCbJX0L+1gxh+WVmWhdxG1IlA\nXZSfCNrVTqhELCv1xcQFNE/uvTKvnotG+rwXarwlQJmsTm2j3FdyL9W6q4ynkkkc\nq+Q4pKX6MsQJUz3Ow9BYBJdwZVozxS9mwLBk2Umzxd7oXSapx03CCQIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFD20\nzN0bwDzmoRhcvMBWQMZSVX6qMB8GA1UdIwQYMBaAFL+dWTj5pA1P7iztXCloltUf\natB6MA0GCSqGSIb3DQEBCwUAA4IBAQAhsxwk88J/9TKaey5JNNF3P6o37RodMNA4\n+9sGKg0Cd8XBFeDRq19cDc0FuMgrEr6NrnX782tlhMV5saeaI1CSZocoITYJUl1z\nSw09m28pigtIr0sUZutbZ/Lv4SDlpYkmNRFyyIMtSvemTt4JZdG1wt5i8IkfuhiW\ns4gKRcCUlIeUG7vxsJ5Ig6sbOJizts8MJ2/Sp4QtfIG5i4LthjNOQgBpNmd8I8wj\nM9432nGktyxlsxOmlTyEAWKPLN29Uab/arPxzACXkwyymqLLFur+RM+dQRNw/PuX\nWFwlo+VAEH9vQsmQYE1XlrxQPARSC6S3xb62MOduIEEzRXKANEd9\n-----END CERTIFICATE-----\n', 'key': '/etc/pki/tls/private/quadlet_demo.key', 'key_content': '-----BEGIN PRIVATE KEY-----\nMIIEvgIBADANBgkqhkiG9w0BAQEFAASCBKgwggSkAgEAAoIBAQDXvmHzmoj7QiuE\nxmQFl5urtCsszLMGBCAuZOaR1eSkAFkb8s0SM66cDJgAGUKGTXE1W9bQcmMVHHOv\nJYgJ13bcargPGCaMaQNlEhIWHUJWdY1TR3XHhPyVhst5HprFqV/5aVYwAkPDrx+1\nK8LR69XwxxDK8HZDL5w7Qt48pSDPneTT/WiR9rijiIwL29F7Cgtx85gupdYJslfQ\nv7WDGH5ZWZaF3EbUiUBdlJ8I2tVOqEQsK/XFxAU0T+69Mq+ei0b6vBdqvCVAmaxO\nbaPcV3Iv1bqrjKeSSRyr5DikpfoyxAlTPc7D0FgEl3BlWjPFL2bAsGTZSbPF3uhd\nJqnHTcIJAgMBAAECggEBAKfNnqY5Bl5Xgz3MiTOITj5o7kP0GzKMtvMe6+92bFXb\nJ3QDj9medStO6FL7UrDfPzR/cD/FGKfs71XFXWA6RGsAXpTu9dy0OT29pm+o5lYV\nLI0fwc4j88nhhBV65mrjaJC3i5/fwxFK/uCNXwSj6E1UqA0Qpwtu1peJVQ3aagDQ\noqKwj/S94i+UMr94qXmWekilihbQClVRTPxwJUo9yAAGP4IlP6zrxz98CGlOO+eD\niEM9yIwrwgNU1cH7V3nCQ1J5KxcikXk80QfJEsDSZ4xP4tt5U+4tPQ0Lr17Rysfn\nuW6eTcgHmScakCyW3qUVmlq6QdJEIg2ztN/kS28Y/lkCgYEA9XcA1HkvUyNEO6pD\n7DfkWlJxeeIzJovXYwM9Nzdg4i9vyj9++3/3MpdEvD+uD5451l0s7/1RB+yRMnh0\nckyNy/tRhKLnvCGeZYTiDvuk8Hr6JVHbkcz+p28vj35MyDzVs3KaIihtm9A0vaX+\n9k161TN4TmqgVg6awZD4DileEH8CgYEA4QDS5rwrwT0ZM7n00i1Cx+sHCpmT6p2l\nv/iLRgFfq67mJ3m3okiHtREYgxEPGF2JzQZmh9E00GL7JYoDbXoy7lVIai4K6Hvz\nUe74THqKY0IsoqRpHtAklas62dotJxqw08gSfcx3Jj052f61+XUnIIVkOD0nPXbf\niNelliAwaXcCgYABnyypOre+U+BSiJh0ShY/JoZLaNwvqALfS5ku8TW5qNiTPIIp\nRTor4mX43G7Y+ZGmQe1XFJRhegWnkLYsPyc8rsdmGfFwA8TZJNydfrojQJzAfF74\nAqO4FXr59rqpJAXBbsNn8F5xBhojGqKpm+9w0ys7UWcHyobpOnz5EvrlHwKBgQCr\nmJI/cBC6jdszjMU+qU/94at8FQYu7s3i6y1kVVXePEPApV48AK6mk4QrKQ5nGacn\nWpNFT3LE3kgFKA1XJ9PB2f6hI74j8avut55bzTTPYD9kRpj46+qKzozSakb3RK7I\nxRP9JrtTyWveahfkSMlvTuvQzCqqAIeX+EaLiL8DHwKBgBd9dR7ZAzzOlHNyv2tq\nfJUP7vka2tUXSvS9hv0owIByfWaPg6EIECcigjn6ljnxIMFJF4UQuq5sIQIppwIx\n6TLLyJqnq8fX4cXhxuMwTvOZnF1j6rsu/RxfhsgK0xUpPknAEOnvdFR/WBdg9O9/\nMPh3fRVIts83uLScKMMMk79S\n-----END PRIVATE KEY-----\n', 'ca': '/etc/pki/tls/certs/quadlet_demo.crt', 'ca_content': '-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRAPbpEtIlyEVcjYU28NLSt60wDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Y2\nZTkxMmQyLTI1Yzg0NTVjLThkODUzNmYwLWQyZDJiN2FiMB4XDTI0MDgwMTAxMjM1\nNVoXDTI1MDgwMTAxMDkxMlowFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA175h85qI+0IrhMZkBZebq7QrLMyzBgQg\nLmTmkdXkpABZG/LNEjOunAyYABlChk1xNVvW0HJjFRxzryWICdd23Gq4DxgmjGkD\nZRISFh1CVnWNU0d1x4T8lYbLeR6axalf+WlWMAJDw68ftSvC0evV8McQyvB2Qy+c\nO0LePKUgz53k0/1okfa4o4iMC9vRewoLcfOYLqXWCbJX0L+1gxh+WVmWhdxG1IlA\nXZSfCNrVTqhELCv1xcQFNE/uvTKvnotG+rwXarwlQJmsTm2j3FdyL9W6q4ynkkkc\nq+Q4pKX6MsQJUz3Ow9BYBJdwZVozxS9mwLBk2Umzxd7oXSapx03CCQIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFD20\nzN0bwDzmoRhcvMBWQMZSVX6qMB8GA1UdIwQYMBaAFL+dWTj5pA1P7iztXCloltUf\natB6MA0GCSqGSIb3DQEBCwUAA4IBAQAhsxwk88J/9TKaey5JNNF3P6o37RodMNA4\n+9sGKg0Cd8XBFeDRq19cDc0FuMgrEr6NrnX782tlhMV5saeaI1CSZocoITYJUl1z\nSw09m28pigtIr0sUZutbZ/Lv4SDlpYkmNRFyyIMtSvemTt4JZdG1wt5i8IkfuhiW\ns4gKRcCUlIeUG7vxsJ5Ig6sbOJizts8MJ2/Sp4QtfIG5i4LthjNOQgBpNmd8I8wj\nM9432nGktyxlsxOmlTyEAWKPLN29Uab/arPxzACXkwyymqLLFur+RM+dQRNw/PuX\nWFwlo+VAEH9vQsmQYE1XlrxQPARSC6S3xb62MOduIEEzRXKANEd9\n-----END CERTIFICATE-----\n'}) => { "ansible_loop_var": "item", "changed": false, "cmd": [ "getcert", "stop-tracking", "-f", "/etc/pki/tls/certs/quadlet_demo.crt" ], "delta": "0:00:00.032145", "end": "2024-07-31 21:23:57.449393", "item": { "ca": "/etc/pki/tls/certs/quadlet_demo.crt", "ca_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRAPbpEtIlyEVcjYU28NLSt60wDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Y2\nZTkxMmQyLTI1Yzg0NTVjLThkODUzNmYwLWQyZDJiN2FiMB4XDTI0MDgwMTAxMjM1\nNVoXDTI1MDgwMTAxMDkxMlowFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA175h85qI+0IrhMZkBZebq7QrLMyzBgQg\nLmTmkdXkpABZG/LNEjOunAyYABlChk1xNVvW0HJjFRxzryWICdd23Gq4DxgmjGkD\nZRISFh1CVnWNU0d1x4T8lYbLeR6axalf+WlWMAJDw68ftSvC0evV8McQyvB2Qy+c\nO0LePKUgz53k0/1okfa4o4iMC9vRewoLcfOYLqXWCbJX0L+1gxh+WVmWhdxG1IlA\nXZSfCNrVTqhELCv1xcQFNE/uvTKvnotG+rwXarwlQJmsTm2j3FdyL9W6q4ynkkkc\nq+Q4pKX6MsQJUz3Ow9BYBJdwZVozxS9mwLBk2Umzxd7oXSapx03CCQIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFD20\nzN0bwDzmoRhcvMBWQMZSVX6qMB8GA1UdIwQYMBaAFL+dWTj5pA1P7iztXCloltUf\natB6MA0GCSqGSIb3DQEBCwUAA4IBAQAhsxwk88J/9TKaey5JNNF3P6o37RodMNA4\n+9sGKg0Cd8XBFeDRq19cDc0FuMgrEr6NrnX782tlhMV5saeaI1CSZocoITYJUl1z\nSw09m28pigtIr0sUZutbZ/Lv4SDlpYkmNRFyyIMtSvemTt4JZdG1wt5i8IkfuhiW\ns4gKRcCUlIeUG7vxsJ5Ig6sbOJizts8MJ2/Sp4QtfIG5i4LthjNOQgBpNmd8I8wj\nM9432nGktyxlsxOmlTyEAWKPLN29Uab/arPxzACXkwyymqLLFur+RM+dQRNw/PuX\nWFwlo+VAEH9vQsmQYE1XlrxQPARSC6S3xb62MOduIEEzRXKANEd9\n-----END CERTIFICATE-----\n", "cert": "/etc/pki/tls/certs/quadlet_demo.crt", "cert_content": "-----BEGIN CERTIFICATE-----\nMIIDgzCCAmugAwIBAgIRAPbpEtIlyEVcjYU28NLSt60wDQYJKoZIhvcNAQELBQAw\nUDEgMB4GA1UEAwwXTG9jYWwgU2lnbmluZyBBdXRob3JpdHkxLDAqBgNVBAMMI2Y2\nZTkxMmQyLTI1Yzg0NTVjLThkODUzNmYwLWQyZDJiN2FiMB4XDTI0MDgwMTAxMjM1\nNVoXDTI1MDgwMTAxMDkxMlowFDESMBAGA1UEAxMJbG9jYWxob3N0MIIBIjANBgkq\nhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA175h85qI+0IrhMZkBZebq7QrLMyzBgQg\nLmTmkdXkpABZG/LNEjOunAyYABlChk1xNVvW0HJjFRxzryWICdd23Gq4DxgmjGkD\nZRISFh1CVnWNU0d1x4T8lYbLeR6axalf+WlWMAJDw68ftSvC0evV8McQyvB2Qy+c\nO0LePKUgz53k0/1okfa4o4iMC9vRewoLcfOYLqXWCbJX0L+1gxh+WVmWhdxG1IlA\nXZSfCNrVTqhELCv1xcQFNE/uvTKvnotG+rwXarwlQJmsTm2j3FdyL9W6q4ynkkkc\nq+Q4pKX6MsQJUz3Ow9BYBJdwZVozxS9mwLBk2Umzxd7oXSapx03CCQIDAQABo4GT\nMIGQMAsGA1UdDwQEAwIFoDAUBgNVHREEDTALgglsb2NhbGhvc3QwHQYDVR0lBBYw\nFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAwHQYDVR0OBBYEFD20\nzN0bwDzmoRhcvMBWQMZSVX6qMB8GA1UdIwQYMBaAFL+dWTj5pA1P7iztXCloltUf\natB6MA0GCSqGSIb3DQEBCwUAA4IBAQAhsxwk88J/9TKaey5JNNF3P6o37RodMNA4\n+9sGKg0Cd8XBFeDRq19cDc0FuMgrEr6NrnX782tlhMV5saeaI1CSZocoITYJUl1z\nSw09m28pigtIr0sUZutbZ/Lv4SDlpYkmNRFyyIMtSvemTt4JZdG1wt5i8IkfuhiW\ns4gKRcCUlIeUG7vxsJ5Ig6sbOJizts8MJ2/Sp4QtfIG5i4LthjNOQgBpNmd8I8wj\nM9432nGktyxlsxOmlTyEAWKPLN29Uab/arPxzACXkwyymqLLFur+RM+dQRNw/PuX\nWFwlo+VAEH9vQsmQYE1XlrxQPARSC6S3xb62MOduIEEzRXKANEd9\n-----END CERTIFICATE-----\n", "key": "/etc/pki/tls/private/quadlet_demo.key", "key_content": "-----BEGIN PRIVATE KEY-----\nMIIEvgIBADANBgkqhkiG9w0BAQEFAASCBKgwggSkAgEAAoIBAQDXvmHzmoj7QiuE\nxmQFl5urtCsszLMGBCAuZOaR1eSkAFkb8s0SM66cDJgAGUKGTXE1W9bQcmMVHHOv\nJYgJ13bcargPGCaMaQNlEhIWHUJWdY1TR3XHhPyVhst5HprFqV/5aVYwAkPDrx+1\nK8LR69XwxxDK8HZDL5w7Qt48pSDPneTT/WiR9rijiIwL29F7Cgtx85gupdYJslfQ\nv7WDGH5ZWZaF3EbUiUBdlJ8I2tVOqEQsK/XFxAU0T+69Mq+ei0b6vBdqvCVAmaxO\nbaPcV3Iv1bqrjKeSSRyr5DikpfoyxAlTPc7D0FgEl3BlWjPFL2bAsGTZSbPF3uhd\nJqnHTcIJAgMBAAECggEBAKfNnqY5Bl5Xgz3MiTOITj5o7kP0GzKMtvMe6+92bFXb\nJ3QDj9medStO6FL7UrDfPzR/cD/FGKfs71XFXWA6RGsAXpTu9dy0OT29pm+o5lYV\nLI0fwc4j88nhhBV65mrjaJC3i5/fwxFK/uCNXwSj6E1UqA0Qpwtu1peJVQ3aagDQ\noqKwj/S94i+UMr94qXmWekilihbQClVRTPxwJUo9yAAGP4IlP6zrxz98CGlOO+eD\niEM9yIwrwgNU1cH7V3nCQ1J5KxcikXk80QfJEsDSZ4xP4tt5U+4tPQ0Lr17Rysfn\nuW6eTcgHmScakCyW3qUVmlq6QdJEIg2ztN/kS28Y/lkCgYEA9XcA1HkvUyNEO6pD\n7DfkWlJxeeIzJovXYwM9Nzdg4i9vyj9++3/3MpdEvD+uD5451l0s7/1RB+yRMnh0\nckyNy/tRhKLnvCGeZYTiDvuk8Hr6JVHbkcz+p28vj35MyDzVs3KaIihtm9A0vaX+\n9k161TN4TmqgVg6awZD4DileEH8CgYEA4QDS5rwrwT0ZM7n00i1Cx+sHCpmT6p2l\nv/iLRgFfq67mJ3m3okiHtREYgxEPGF2JzQZmh9E00GL7JYoDbXoy7lVIai4K6Hvz\nUe74THqKY0IsoqRpHtAklas62dotJxqw08gSfcx3Jj052f61+XUnIIVkOD0nPXbf\niNelliAwaXcCgYABnyypOre+U+BSiJh0ShY/JoZLaNwvqALfS5ku8TW5qNiTPIIp\nRTor4mX43G7Y+ZGmQe1XFJRhegWnkLYsPyc8rsdmGfFwA8TZJNydfrojQJzAfF74\nAqO4FXr59rqpJAXBbsNn8F5xBhojGqKpm+9w0ys7UWcHyobpOnz5EvrlHwKBgQCr\nmJI/cBC6jdszjMU+qU/94at8FQYu7s3i6y1kVVXePEPApV48AK6mk4QrKQ5nGacn\nWpNFT3LE3kgFKA1XJ9PB2f6hI74j8avut55bzTTPYD9kRpj46+qKzozSakb3RK7I\nxRP9JrtTyWveahfkSMlvTuvQzCqqAIeX+EaLiL8DHwKBgBd9dR7ZAzzOlHNyv2tq\nfJUP7vka2tUXSvS9hv0owIByfWaPg6EIECcigjn6ljnxIMFJF4UQuq5sIQIppwIx\n6TLLyJqnq8fX4cXhxuMwTvOZnF1j6rsu/RxfhsgK0xUpPknAEOnvdFR/WBdg9O9/\nMPh3fRVIts83uLScKMMMk79S\n-----END PRIVATE KEY-----\n" }, "rc": 0, "start": "2024-07-31 21:23:57.417248" } STDOUT: Request "20240801012355" removed. TASK [fedora.linux_system_roles.certificate : Remove files] ******************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:181 Wednesday 31 July 2024 21:23:57 -0400 (0:00:00.478) 0:00:10.943 ******** changed: [managed_node1] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } changed: [managed_node1] => (item=/etc/pki/tls/private/quadlet_demo.key) => { "ansible_loop_var": "item", "changed": true, "item": "/etc/pki/tls/private/quadlet_demo.key", "path": "/etc/pki/tls/private/quadlet_demo.key", "state": "absent" } ok: [managed_node1] => (item=/etc/pki/tls/certs/quadlet_demo.crt) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/pki/tls/certs/quadlet_demo.crt", "path": "/etc/pki/tls/certs/quadlet_demo.crt", "state": "absent" } TASK [Run the role] ************************************************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:50 Wednesday 31 July 2024 21:23:58 -0400 (0:00:01.011) 0:00:11.954 ******** TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Wednesday 31 July 2024 21:23:58 -0400 (0:00:00.071) 0:00:12.025 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Wednesday 31 July 2024 21:23:58 -0400 (0:00:00.024) 0:00:12.050 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Wednesday 31 July 2024 21:23:58 -0400 (0:00:00.032) 0:00:12.083 ******** ok: [managed_node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Wednesday 31 July 2024 21:23:58 -0400 (0:00:00.343) 0:00:12.426 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:20 Wednesday 31 July 2024 21:23:59 -0400 (0:00:00.036) 0:00:12.463 ******** ok: [managed_node1] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed_node1] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node1] => (item=CentOS_8.yml) => { "ansible_facts": { "__podman_packages": [ "crun", "podman", "podman-plugins", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml" } ok: [managed_node1] => (item=CentOS_8.yml) => { "ansible_facts": { "__podman_packages": [ "crun", "podman", "podman-plugins", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Wednesday 31 July 2024 21:23:59 -0400 (0:00:00.058) 0:00:12.522 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Wednesday 31 July 2024 21:24:00 -0400 (0:00:01.626) 0:00:14.148 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Wednesday 31 July 2024 21:24:00 -0400 (0:00:00.048) 0:00:14.197 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:22 Wednesday 31 July 2024 21:24:00 -0400 (0:00:00.054) 0:00:14.251 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.029410", "end": "2024-07-31 21:24:01.117960", "rc": 0, "start": "2024-07-31 21:24:01.088550" } STDOUT: podman version 4.9.4-dev TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Wednesday 31 July 2024 21:24:01 -0400 (0:00:00.372) 0:00:14.623 ******** ok: [managed_node1] => { "ansible_facts": { "podman_version": "4.9.4-dev" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:32 Wednesday 31 July 2024 21:24:01 -0400 (0:00:00.048) 0:00:14.672 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:39 Wednesday 31 July 2024 21:24:01 -0400 (0:00:00.047) 0:00:14.720 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } META: end_host conditional evaluated to false, continuing execution for managed_node1 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Wednesday 31 July 2024 21:24:01 -0400 (0:00:00.172) 0:00:14.892 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:24:01 -0400 (0:00:00.058) 0:00:14.951 ******** ok: [managed_node1] => { "ansible_facts": { "getent_passwd": { "root": [ "x", "0", "0", "root", "/root", "/bin/bash" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:24:01 -0400 (0:00:00.471) 0:00:15.422 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:24:02 -0400 (0:00:00.034) 0:00:15.456 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:24:02 -0400 (0:00:00.063) 0:00:15.520 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:24:02 -0400 (0:00:00.365) 0:00:15.885 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:24:02 -0400 (0:00:00.037) 0:00:15.923 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:24:02 -0400 (0:00:00.358) 0:00:16.281 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:24:02 -0400 (0:00:00.030) 0:00:16.312 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:24:02 -0400 (0:00:00.027) 0:00:16.339 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:24:02 -0400 (0:00:00.029) 0:00:16.369 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:24:02 -0400 (0:00:00.027) 0:00:16.396 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:24:02 -0400 (0:00:00.027) 0:00:16.424 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.027) 0:00:16.451 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.029) 0:00:16.481 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:62 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.027) 0:00:16.508 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:71 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.059) 0:00:16.568 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.055) 0:00:16.623 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.029) 0:00:16.653 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:74 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.057) 0:00:16.710 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.057) 0:00:16.767 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.027) 0:00:16.795 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:77 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.028) 0:00:16.824 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.056) 0:00:16.881 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.027) 0:00:16.909 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.027) 0:00:16.937 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.056) 0:00:16.994 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.029) 0:00:17.024 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.027) 0:00:17.051 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.028) 0:00:17.079 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:86 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.027) 0:00:17.107 ******** TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.120) 0:00:17.227 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed_node1 TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.055) 0:00:17.283 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Wednesday 31 July 2024 21:24:03 -0400 (0:00:00.036) 0:00:17.319 ******** ok: [managed_node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Wednesday 31 July 2024 21:24:04 -0400 (0:00:00.349) 0:00:17.669 ******** ok: [managed_node1] => { "ansible_facts": { "__firewall_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22 Wednesday 31 July 2024 21:24:04 -0400 (0:00:00.033) 0:00:17.703 ******** ok: [managed_node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27 Wednesday 31 July 2024 21:24:04 -0400 (0:00:00.344) 0:00:18.047 ******** ok: [managed_node1] => { "ansible_facts": { "__firewall_is_transactional": false }, "changed": false } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 Wednesday 31 July 2024 21:24:04 -0400 (0:00:00.032) 0:00:18.080 ******** ok: [managed_node1] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43 Wednesday 31 July 2024 21:24:07 -0400 (0:00:02.414) 0:00:20.494 ******** skipping: [managed_node1] => {} TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48 Wednesday 31 July 2024 21:24:07 -0400 (0:00:00.030) 0:00:20.524 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53 Wednesday 31 July 2024 21:24:07 -0400 (0:00:00.030) 0:00:20.555 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Wednesday 31 July 2024 21:24:07 -0400 (0:00:00.027) 0:00:20.582 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Wednesday 31 July 2024 21:24:07 -0400 (0:00:00.030) 0:00:20.612 ******** skipping: [managed_node1] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "item": "ufw", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Wednesday 31 July 2024 21:24:07 -0400 (0:00:00.036) 0:00:20.648 ******** ok: [managed_node1] => { "changed": false, "name": "firewalld", "status": { "ActiveEnterTimestamp": "Wed 2024-07-31 21:19:43 EDT", "ActiveEnterTimestampMonotonic": "6479081", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "polkit.service sysinit.target system.slice dbus.service basic.target dbus.socket", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-31 21:19:42 EDT", "AssertTimestampMonotonic": "5431503", "Before": "network-pre.target multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-31 21:19:42 EDT", "ConditionTimestampMonotonic": "5431502", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target nftables.service iptables.service ip6tables.service ebtables.service ipset.service", "ControlGroup": "/system.slice/firewalld.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "man:firewalld(1)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "582", "ExecMainStartTimestamp": "Wed 2024-07-31 21:19:42 EDT", "ExecMainStartTimestampMonotonic": "5433279", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-31 21:19:42 EDT", "InactiveExitTimestampMonotonic": "5433306", "InvocationID": "a1cc5d84d60d46859a9019e5cb14bc27", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "582", "MemoryAccounting": "yes", "MemoryCurrent": "53256192", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "dbus-org.fedoraproject.FirewallD1.service firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target dbus.socket system.slice", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Wed 2024-07-31 21:19:43 EDT", "StateChangeTimestampMonotonic": "6479081", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogTimestamp": "Wed 2024-07-31 21:19:43 EDT", "WatchdogTimestampMonotonic": "6479078", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Wednesday 31 July 2024 21:24:07 -0400 (0:00:00.493) 0:00:21.142 ******** ok: [managed_node1] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "ActiveEnterTimestamp": "Wed 2024-07-31 21:19:43 EDT", "ActiveEnterTimestampMonotonic": "6479081", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "polkit.service sysinit.target system.slice dbus.service basic.target dbus.socket", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-31 21:19:42 EDT", "AssertTimestampMonotonic": "5431503", "Before": "network-pre.target multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-31 21:19:42 EDT", "ConditionTimestampMonotonic": "5431502", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target nftables.service iptables.service ip6tables.service ebtables.service ipset.service", "ControlGroup": "/system.slice/firewalld.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "man:firewalld(1)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "582", "ExecMainStartTimestamp": "Wed 2024-07-31 21:19:42 EDT", "ExecMainStartTimestampMonotonic": "5433279", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-31 21:19:42 EDT", "InactiveExitTimestampMonotonic": "5433306", "InvocationID": "a1cc5d84d60d46859a9019e5cb14bc27", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "582", "MemoryAccounting": "yes", "MemoryCurrent": "53256192", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "dbus-org.fedoraproject.FirewallD1.service firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target dbus.socket system.slice", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Wed 2024-07-31 21:19:43 EDT", "StateChangeTimestampMonotonic": "6479081", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogTimestamp": "Wed 2024-07-31 21:19:43 EDT", "WatchdogTimestampMonotonic": "6479078", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Wednesday 31 July 2024 21:24:08 -0400 (0:00:00.477) 0:00:21.620 ******** ok: [managed_node1] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/libexec/platform-python", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Wednesday 31 July 2024 21:24:08 -0400 (0:00:00.036) 0:00:21.656 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Wednesday 31 July 2024 21:24:08 -0400 (0:00:00.031) 0:00:21.688 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Wednesday 31 July 2024 21:24:08 -0400 (0:00:00.031) 0:00:21.719 ******** changed: [managed_node1] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "8000/tcp", "state": "enabled" } } changed: [managed_node1] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": true, "ansible_loop_var": "item", "changed": true, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Wednesday 31 July 2024 21:24:09 -0400 (0:00:01.206) 0:00:22.925 ******** skipping: [managed_node1] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Wednesday 31 July 2024 21:24:09 -0400 (0:00:00.049) 0:00:22.974 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Wednesday 31 July 2024 21:24:09 -0400 (0:00:00.032) 0:00:23.007 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Wednesday 31 July 2024 21:24:09 -0400 (0:00:00.031) 0:00:23.038 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Wednesday 31 July 2024 21:24:09 -0400 (0:00:00.029) 0:00:23.067 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Wednesday 31 July 2024 21:24:09 -0400 (0:00:00.029) 0:00:23.096 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Wednesday 31 July 2024 21:24:09 -0400 (0:00:00.027) 0:00:23.124 ******** skipping: [managed_node1] => {} TASK [Manage selinux for specified ports] ************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:93 Wednesday 31 July 2024 21:24:09 -0400 (0:00:00.029) 0:00:23.154 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:100 Wednesday 31 July 2024 21:24:09 -0400 (0:00:00.026) 0:00:23.181 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:104 Wednesday 31 July 2024 21:24:09 -0400 (0:00:00.030) 0:00:23.211 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:113 Wednesday 31 July 2024 21:24:09 -0400 (0:00:00.025) 0:00:23.236 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:122 Wednesday 31 July 2024 21:24:09 -0400 (0:00:00.027) 0:00:23.264 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Wednesday 31 July 2024 21:24:09 -0400 (0:00:00.153) 0:00:23.418 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Wednesday 31 July 2024 21:24:10 -0400 (0:00:00.033) 0:00:23.451 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Wednesday 31 July 2024 21:24:10 -0400 (0:00:00.040) 0:00:23.491 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:24:10 -0400 (0:00:00.045) 0:00:23.537 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:24:10 -0400 (0:00:00.030) 0:00:23.568 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:24:10 -0400 (0:00:00.027) 0:00:23.596 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Wednesday 31 July 2024 21:24:10 -0400 (0:00:00.032) 0:00:23.628 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Wednesday 31 July 2024 21:24:10 -0400 (0:00:00.028) 0:00:23.657 ******** [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Wednesday 31 July 2024 21:24:10 -0400 (0:00:00.618) 0:00:24.275 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Wednesday 31 July 2024 21:24:10 -0400 (0:00:00.032) 0:00:24.308 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Wednesday 31 July 2024 21:24:10 -0400 (0:00:00.038) 0:00:24.346 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:24:10 -0400 (0:00:00.046) 0:00:24.392 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:24:10 -0400 (0:00:00.027) 0:00:24.419 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:24:11 -0400 (0:00:00.030) 0:00:24.450 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Wednesday 31 July 2024 21:24:11 -0400 (0:00:00.056) 0:00:24.506 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Wednesday 31 July 2024 21:24:11 -0400 (0:00:00.028) 0:00:24.535 ******** [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Wednesday 31 July 2024 21:24:11 -0400 (0:00:00.490) 0:00:25.026 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Wednesday 31 July 2024 21:24:11 -0400 (0:00:00.032) 0:00:25.059 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Wednesday 31 July 2024 21:24:11 -0400 (0:00:00.039) 0:00:25.098 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:24:11 -0400 (0:00:00.047) 0:00:25.145 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:24:11 -0400 (0:00:00.028) 0:00:25.173 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:24:11 -0400 (0:00:00.029) 0:00:25.203 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Wednesday 31 July 2024 21:24:11 -0400 (0:00:00.026) 0:00:25.230 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Wednesday 31 July 2024 21:24:11 -0400 (0:00:00.029) 0:00:25.260 ******** [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:129 Wednesday 31 July 2024 21:24:12 -0400 (0:00:00.489) 0:00:25.749 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:136 Wednesday 31 July 2024 21:24:12 -0400 (0:00:00.026) 0:00:25.776 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 31 July 2024 21:24:12 -0400 (0:00:00.154) 0:00:25.930 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 31 July 2024 21:24:12 -0400 (0:00:00.042) 0:00:25.973 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 31 July 2024 21:24:12 -0400 (0:00:00.035) 0:00:26.009 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 31 July 2024 21:24:12 -0400 (0:00:00.032) 0:00:26.041 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 31 July 2024 21:24:12 -0400 (0:00:00.040) 0:00:26.081 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:24:12 -0400 (0:00:00.054) 0:00:26.136 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:24:12 -0400 (0:00:00.038) 0:00:26.174 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:24:12 -0400 (0:00:00.035) 0:00:26.209 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:24:12 -0400 (0:00:00.038) 0:00:26.247 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.360) 0:00:26.608 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.036) 0:00:26.644 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.346) 0:00:26.991 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.028) 0:00:27.019 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.027) 0:00:27.047 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.061) 0:00:27.108 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.028) 0:00:27.137 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.030) 0:00:27.168 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.029) 0:00:27.197 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.029) 0:00:27.226 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.029) 0:00:27.255 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-network.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.052) 0:00:27.308 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.032) 0:00:27.340 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.029) 0:00:27.370 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 31 July 2024 21:24:13 -0400 (0:00:00.066) 0:00:27.436 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 31 July 2024 21:24:14 -0400 (0:00:00.035) 0:00:27.472 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 31 July 2024 21:24:14 -0400 (0:00:00.029) 0:00:27.501 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Wednesday 31 July 2024 21:24:14 -0400 (0:00:00.065) 0:00:27.567 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:24:14 -0400 (0:00:00.050) 0:00:27.617 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:24:14 -0400 (0:00:00.028) 0:00:27.646 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:24:14 -0400 (0:00:00.027) 0:00:27.673 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Wednesday 31 July 2024 21:24:14 -0400 (0:00:00.063) 0:00:27.737 ******** TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Wednesday 31 July 2024 21:24:14 -0400 (0:00:00.027) 0:00:27.764 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Wednesday 31 July 2024 21:24:14 -0400 (0:00:00.027) 0:00:27.792 ******** ok: [managed_node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Wednesday 31 July 2024 21:24:14 -0400 (0:00:00.365) 0:00:28.158 ******** changed: [managed_node1] => { "changed": true, "checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0", "dest": "/etc/containers/systemd/quadlet-demo.network", "gid": 0, "group": "root", "md5sum": "061f3cf318cbd8ab5794bb1173831fb8", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 74, "src": "/root/.ansible/tmp/ansible-tmp-1722475454.7581224-28360-69997062430020/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Wednesday 31 July 2024 21:24:15 -0400 (0:00:00.729) 0:00:28.887 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Wednesday 31 July 2024 21:24:15 -0400 (0:00:00.030) 0:00:28.918 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Wednesday 31 July 2024 21:24:15 -0400 (0:00:00.030) 0:00:28.948 ******** ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Wednesday 31 July 2024 21:24:16 -0400 (0:00:00.589) 0:00:29.538 ******** changed: [managed_node1] => { "changed": true, "name": "quadlet-demo-network.service", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "basic.target systemd-journald.socket system.slice -.mount sysinit.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-network.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-network.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-network.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-network.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "system.slice -.mount sysinit.target", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-network", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22405", "TimeoutStartUSec": "infinity", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Wednesday 31 July 2024 21:24:16 -0400 (0:00:00.560) 0:00:30.099 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 31 July 2024 21:24:16 -0400 (0:00:00.063) 0:00:30.162 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 31 July 2024 21:24:16 -0400 (0:00:00.039) 0:00:30.202 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 31 July 2024 21:24:16 -0400 (0:00:00.037) 0:00:30.240 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 31 July 2024 21:24:16 -0400 (0:00:00.029) 0:00:30.269 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 31 July 2024 21:24:16 -0400 (0:00:00.044) 0:00:30.313 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:24:16 -0400 (0:00:00.055) 0:00:30.369 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:24:16 -0400 (0:00:00.037) 0:00:30.406 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:24:17 -0400 (0:00:00.034) 0:00:30.440 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:24:17 -0400 (0:00:00.040) 0:00:30.481 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:24:17 -0400 (0:00:00.368) 0:00:30.849 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:24:17 -0400 (0:00:00.037) 0:00:30.887 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:24:17 -0400 (0:00:00.350) 0:00:31.238 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:24:17 -0400 (0:00:00.030) 0:00:31.268 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:24:17 -0400 (0:00:00.031) 0:00:31.299 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:24:17 -0400 (0:00:00.029) 0:00:31.329 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:24:17 -0400 (0:00:00.068) 0:00:31.397 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:24:17 -0400 (0:00:00.029) 0:00:31.426 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.030) 0:00:31.457 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.028) 0:00:31.485 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.029) 0:00:31.514 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.049) 0:00:31.564 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.030) 0:00:31.595 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.029) 0:00:31.624 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.070) 0:00:31.695 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.032) 0:00:31.728 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.030) 0:00:31.758 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.062) 0:00:31.820 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.050) 0:00:31.871 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.028) 0:00:31.899 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.030) 0:00:31.929 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.026) 0:00:31.956 ******** TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.061) 0:00:32.018 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.029) 0:00:32.048 ******** ok: [managed_node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 34, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Wednesday 31 July 2024 21:24:18 -0400 (0:00:00.365) 0:00:32.413 ******** changed: [managed_node1] => { "changed": true, "checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a", "dest": "/etc/containers/systemd/quadlet-demo-mysql.volume", "gid": 0, "group": "root", "md5sum": "5ddd03a022aeb4502d9bc8ce436b4233", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 9, "src": "/root/.ansible/tmp/ansible-tmp-1722475459.0135796-28487-68246258164666/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Wednesday 31 July 2024 21:24:19 -0400 (0:00:00.658) 0:00:33.071 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Wednesday 31 July 2024 21:24:19 -0400 (0:00:00.030) 0:00:33.102 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Wednesday 31 July 2024 21:24:19 -0400 (0:00:00.030) 0:00:33.132 ******** ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Wednesday 31 July 2024 21:24:20 -0400 (0:00:00.577) 0:00:33.710 ******** changed: [managed_node1] => { "changed": true, "name": "quadlet-demo-mysql-volume.service", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "basic.target -.mount systemd-journald.socket sysinit.target system.slice", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "-.mount sysinit.target system.slice", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22405", "TimeoutStartUSec": "infinity", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Wednesday 31 July 2024 21:24:20 -0400 (0:00:00.537) 0:00:34.247 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 31 July 2024 21:24:20 -0400 (0:00:00.033) 0:00:34.281 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 31 July 2024 21:24:21 -0400 (0:00:00.164) 0:00:34.445 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 31 July 2024 21:24:21 -0400 (0:00:00.040) 0:00:34.486 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 31 July 2024 21:24:21 -0400 (0:00:00.034) 0:00:34.520 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 31 July 2024 21:24:21 -0400 (0:00:00.042) 0:00:34.562 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:24:21 -0400 (0:00:00.057) 0:00:34.620 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:24:21 -0400 (0:00:00.037) 0:00:34.657 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:24:21 -0400 (0:00:00.035) 0:00:34.693 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:24:21 -0400 (0:00:00.042) 0:00:34.736 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:24:21 -0400 (0:00:00.371) 0:00:35.108 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:24:21 -0400 (0:00:00.038) 0:00:35.146 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.352) 0:00:35.499 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.030) 0:00:35.530 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.028) 0:00:35.558 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.028) 0:00:35.587 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.027) 0:00:35.615 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.060) 0:00:35.676 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.029) 0:00:35.705 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.028) 0:00:35.734 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.028) 0:00:35.763 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.051) 0:00:35.814 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.031) 0:00:35.846 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.031) 0:00:35.877 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.068) 0:00:35.946 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.033) 0:00:35.979 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.028) 0:00:36.007 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.061) 0:00:36.069 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.049) 0:00:36.118 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.027) 0:00:36.146 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.029) 0:00:36.176 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Wednesday 31 July 2024 21:24:22 -0400 (0:00:00.027) 0:00:36.203 ******** [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => (item=/tmp/quadlet_demo) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/quadlet_demo", "mode": "0777", "owner": "root", "path": "/tmp/quadlet_demo", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Wednesday 31 July 2024 21:24:23 -0400 (0:00:00.416) 0:00:36.620 ******** changed: [managed_node1] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Wednesday 31 July 2024 21:24:30 -0400 (0:00:07.148) 0:00:43.769 ******** ok: [managed_node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 67, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Wednesday 31 July 2024 21:24:30 -0400 (0:00:00.363) 0:00:44.132 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Wednesday 31 July 2024 21:24:30 -0400 (0:00:00.031) 0:00:44.164 ******** changed: [managed_node1] => { "changed": true, "checksum": "ca62b2ad3cc9afb5b5371ebbf797b9bc4fd7edd4", "dest": "/etc/containers/systemd/quadlet-demo-mysql.container", "gid": 0, "group": "root", "md5sum": "341b473056d2a5dfa35970b0d2e23a5d", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 363, "src": "/root/.ansible/tmp/ansible-tmp-1722475470.7698424-28631-129284827082904/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Wednesday 31 July 2024 21:24:31 -0400 (0:00:00.657) 0:00:44.822 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Wednesday 31 July 2024 21:24:31 -0400 (0:00:00.031) 0:00:44.853 ******** ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Wednesday 31 July 2024 21:24:32 -0400 (0:00:00.588) 0:00:45.441 ******** changed: [managed_node1] => { "changed": true, "name": "quadlet-demo-mysql.service", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "system.slice -.mount quadlet-demo-network.service systemd-journald.socket quadlet-demo-mysql-volume.service sysinit.target basic.target tmp.mount", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "yes", "DelegateControllers": "cpu cpuacct cpuset io blkio memory devices pids", "Description": "quadlet-demo-mysql.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-mysql.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "quadlet-demo-mysql-volume.service system.slice quadlet-demo-network.service sysinit.target -.mount", "RequiresMountsFor": "/run/containers /tmp/quadlet_demo", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Wednesday 31 July 2024 21:24:32 -0400 (0:00:00.954) 0:00:46.396 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 31 July 2024 21:24:32 -0400 (0:00:00.031) 0:00:46.428 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "envoy-proxy-configmap.yml", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 31 July 2024 21:24:33 -0400 (0:00:00.040) 0:00:46.468 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 31 July 2024 21:24:33 -0400 (0:00:00.035) 0:00:46.503 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 31 July 2024 21:24:33 -0400 (0:00:00.030) 0:00:46.534 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "envoy-proxy-configmap", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 31 July 2024 21:24:33 -0400 (0:00:00.043) 0:00:46.577 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:24:33 -0400 (0:00:00.055) 0:00:46.632 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:24:33 -0400 (0:00:00.033) 0:00:46.666 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:24:33 -0400 (0:00:00.038) 0:00:46.704 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:24:33 -0400 (0:00:00.040) 0:00:46.745 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:24:33 -0400 (0:00:00.418) 0:00:47.163 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:24:33 -0400 (0:00:00.040) 0:00:47.203 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.378) 0:00:47.582 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.029) 0:00:47.612 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.027) 0:00:47.640 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.028) 0:00:47.668 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.028) 0:00:47.697 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.027) 0:00:47.724 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.030) 0:00:47.754 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.026) 0:00:47.781 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.030) 0:00:47.812 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.051) 0:00:47.863 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.029) 0:00:47.893 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.031) 0:00:47.924 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.068) 0:00:47.992 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.065) 0:00:48.058 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.028) 0:00:48.087 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.065) 0:00:48.152 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.047) 0:00:48.200 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.029) 0:00:48.229 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.028) 0:00:48.257 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.030) 0:00:48.287 ******** TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.025) 0:00:48.313 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Wednesday 31 July 2024 21:24:34 -0400 (0:00:00.028) 0:00:48.342 ******** ok: [managed_node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 103, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Wednesday 31 July 2024 21:24:35 -0400 (0:00:00.362) 0:00:48.705 ******** changed: [managed_node1] => { "changed": true, "checksum": "d681c7d56f912150d041873e880818b22a90c188", "dest": "/etc/containers/systemd/envoy-proxy-configmap.yml", "gid": 0, "group": "root", "md5sum": "aec75d972c231aac004e1338934544cf", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 2102, "src": "/root/.ansible/tmp/ansible-tmp-1722475475.3095858-28756-53264211105254/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Wednesday 31 July 2024 21:24:35 -0400 (0:00:00.705) 0:00:49.411 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Wednesday 31 July 2024 21:24:36 -0400 (0:00:00.031) 0:00:49.442 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Wednesday 31 July 2024 21:24:36 -0400 (0:00:00.067) 0:00:49.510 ******** ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Wednesday 31 July 2024 21:24:36 -0400 (0:00:00.601) 0:00:50.112 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Wednesday 31 July 2024 21:24:36 -0400 (0:00:00.034) 0:00:50.146 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 31 July 2024 21:24:36 -0400 (0:00:00.038) 0:00:50.185 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n", "__podman_quadlet_template_src": "quadlet-demo.yml.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 31 July 2024 21:24:36 -0400 (0:00:00.087) 0:00:50.272 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 31 July 2024 21:24:36 -0400 (0:00:00.037) 0:00:50.309 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 31 July 2024 21:24:36 -0400 (0:00:00.033) 0:00:50.343 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 31 July 2024 21:24:36 -0400 (0:00:00.040) 0:00:50.384 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:24:37 -0400 (0:00:00.056) 0:00:50.441 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:24:37 -0400 (0:00:00.036) 0:00:50.478 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:24:37 -0400 (0:00:00.037) 0:00:50.515 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:24:37 -0400 (0:00:00.038) 0:00:50.554 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:24:37 -0400 (0:00:00.357) 0:00:50.912 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:24:37 -0400 (0:00:00.073) 0:00:50.986 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.504) 0:00:51.490 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.031) 0:00:51.522 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.029) 0:00:51.551 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.031) 0:00:51.583 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.029) 0:00:51.613 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.029) 0:00:51.642 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.028) 0:00:51.671 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.030) 0:00:51.702 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.027) 0:00:51.729 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.052) 0:00:51.781 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.031) 0:00:51.813 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.029) 0:00:51.842 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.068) 0:00:51.910 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.035) 0:00:51.946 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.028) 0:00:51.974 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.094) 0:00:52.068 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.050) 0:00:52.119 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.028) 0:00:52.147 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.029) 0:00:52.177 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.027) 0:00:52.205 ******** TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.028) 0:00:52.234 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Wednesday 31 July 2024 21:24:38 -0400 (0:00:00.026) 0:00:52.260 ******** ok: [managed_node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 136, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Wednesday 31 July 2024 21:24:39 -0400 (0:00:00.494) 0:00:52.754 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Wednesday 31 July 2024 21:24:39 -0400 (0:00:00.030) 0:00:52.784 ******** changed: [managed_node1] => { "changed": true, "checksum": "998dccde0483b1654327a46ddd89cbaa47650370", "dest": "/etc/containers/systemd/quadlet-demo.yml", "gid": 0, "group": "root", "md5sum": "fd890594adfc24339cb9cdc5e7b19a66", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 1605, "src": "/root/.ansible/tmp/ansible-tmp-1722475479.3956273-28878-108312669605419/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Wednesday 31 July 2024 21:24:40 -0400 (0:00:00.945) 0:00:53.730 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Wednesday 31 July 2024 21:24:40 -0400 (0:00:00.030) 0:00:53.761 ******** ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Wednesday 31 July 2024 21:24:40 -0400 (0:00:00.657) 0:00:54.418 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Wednesday 31 July 2024 21:24:41 -0400 (0:00:00.034) 0:00:54.452 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 31 July 2024 21:24:41 -0400 (0:00:00.032) 0:00:54.485 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 31 July 2024 21:24:41 -0400 (0:00:00.042) 0:00:54.528 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 31 July 2024 21:24:41 -0400 (0:00:00.036) 0:00:54.564 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 31 July 2024 21:24:41 -0400 (0:00:00.031) 0:00:54.595 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 31 July 2024 21:24:41 -0400 (0:00:00.040) 0:00:54.636 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:24:41 -0400 (0:00:00.055) 0:00:54.692 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:24:41 -0400 (0:00:00.037) 0:00:54.729 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:24:41 -0400 (0:00:00.036) 0:00:54.765 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:24:41 -0400 (0:00:00.039) 0:00:54.805 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:24:41 -0400 (0:00:00.359) 0:00:55.165 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:24:41 -0400 (0:00:00.038) 0:00:55.203 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:24:42 -0400 (0:00:00.355) 0:00:55.559 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:24:42 -0400 (0:00:00.030) 0:00:55.589 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:24:42 -0400 (0:00:00.027) 0:00:55.617 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:24:42 -0400 (0:00:00.029) 0:00:55.647 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:24:42 -0400 (0:00:00.027) 0:00:55.675 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:24:42 -0400 (0:00:00.029) 0:00:55.704 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:24:42 -0400 (0:00:00.028) 0:00:55.732 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:24:42 -0400 (0:00:00.029) 0:00:55.762 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 31 July 2024 21:24:42 -0400 (0:00:00.027) 0:00:55.789 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": [ "quadlet-demo.yml" ], "__podman_service_name": "quadlet-demo.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 31 July 2024 21:24:42 -0400 (0:00:00.052) 0:00:55.842 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 31 July 2024 21:24:42 -0400 (0:00:00.029) 0:00:55.871 ******** ok: [managed_node1] => { "changed": false, "content": "LS0tCmFwaVZlcnNpb246IHYxCmtpbmQ6IFBlcnNpc3RlbnRWb2x1bWVDbGFpbQptZXRhZGF0YToKICBuYW1lOiB3cC1wdi1jbGFpbQogIGxhYmVsczoKICAgIGFwcDogd29yZHByZXNzCnNwZWM6CiAgYWNjZXNzTW9kZXM6CiAgLSBSZWFkV3JpdGVPbmNlCiAgcmVzb3VyY2VzOgogICAgcmVxdWVzdHM6CiAgICAgIHN0b3JhZ2U6IDIwR2kKLS0tCmFwaVZlcnNpb246IHYxCmtpbmQ6IFBvZAptZXRhZGF0YToKICBuYW1lOiBxdWFkbGV0LWRlbW8Kc3BlYzoKICBjb250YWluZXJzOgogIC0gbmFtZTogd29yZHByZXNzCiAgICBpbWFnZTogcXVheS5pby9saW51eC1zeXN0ZW0tcm9sZXMvd29yZHByZXNzOjQuOC1hcGFjaGUKICAgIGVudjoKICAgIC0gbmFtZTogV09SRFBSRVNTX0RCX0hPU1QKICAgICAgdmFsdWU6IHF1YWRsZXQtZGVtby1teXNxbAogICAgLSBuYW1lOiBXT1JEUFJFU1NfREJfUEFTU1dPUkQKICAgICAgdmFsdWVGcm9tOgogICAgICAgIHNlY3JldEtleVJlZjoKICAgICAgICAgIG5hbWU6IG15c3FsLXJvb3QtcGFzc3dvcmQta3ViZQogICAgICAgICAga2V5OiBwYXNzd29yZAogICAgdm9sdW1lTW91bnRzOgogICAgLSBuYW1lOiB3b3JkcHJlc3MtcGVyc2lzdGVudC1zdG9yYWdlCiAgICAgIG1vdW50UGF0aDogL3Zhci93d3cvaHRtbAogICAgcmVzb3VyY2VzOgogICAgICByZXF1ZXN0czoKICAgICAgICBtZW1vcnk6ICI2NE1pIgogICAgICAgIGNwdTogIjI1MG0iCiAgICAgIGxpbWl0czoKICAgICAgICBtZW1vcnk6ICIxMjhNaSIKICAgICAgICBjcHU6ICI1MDBtIgogIC0gbmFtZTogZW52b3kKICAgIGltYWdlOiBxdWF5LmlvL2xpbnV4LXN5c3RlbS1yb2xlcy9lbnZveXByb3h5OnYxLjI1LjAKICAgIHZvbHVtZU1vdW50czoKICAgIC0gbmFtZTogY29uZmlnLXZvbHVtZQogICAgICBtb3VudFBhdGg6IC9ldGMvZW52b3kKICAgIC0gbmFtZTogY2VydGlmaWNhdGVzCiAgICAgIG1vdW50UGF0aDogL2V0Yy9lbnZveS1jZXJ0aWZpY2F0ZXMKICAgIGVudjoKICAgIC0gbmFtZTogRU5WT1lfVUlECiAgICAgIHZhbHVlOiAiMCIKICAgIHJlc291cmNlczoKICAgICAgcmVxdWVzdHM6CiAgICAgICAgbWVtb3J5OiAiNjRNaSIKICAgICAgICBjcHU6ICIyNTBtIgogICAgICBsaW1pdHM6CiAgICAgICAgbWVtb3J5OiAiMTI4TWkiCiAgICAgICAgY3B1OiAiNTAwbSIKICB2b2x1bWVzOgogIC0gbmFtZTogY29uZmlnLXZvbHVtZQogICAgY29uZmlnTWFwOgogICAgICBuYW1lOiBlbnZveS1wcm94eS1jb25maWcKICAtIG5hbWU6IGNlcnRpZmljYXRlcwogICAgc2VjcmV0OgogICAgICBzZWNyZXROYW1lOiBlbnZveS1jZXJ0aWZpY2F0ZXMKICAtIG5hbWU6IHdvcmRwcmVzcy1wZXJzaXN0ZW50LXN0b3JhZ2UKICAgIHBlcnNpc3RlbnRWb2x1bWVDbGFpbToKICAgICAgY2xhaW1OYW1lOiB3cC1wdi1jbGFpbQogIC0gbmFtZTogd3d3ICAjIG5vdCB1c2VkIC0gZm9yIHRlc3RpbmcgaG9zdHBhdGgKICAgIGhvc3RQYXRoOgogICAgICBwYXRoOiAvdG1wL2h0dHBkMwogIC0gbmFtZTogY3JlYXRlICAjIG5vdCB1c2VkIC0gZm9yIHRlc3RpbmcgaG9zdHBhdGgKICAgIGhvc3RQYXRoOgogICAgICBwYXRoOiAvdG1wL2h0dHBkMy1jcmVhdGUK", "encoding": "base64", "source": "/etc/containers/systemd/quadlet-demo.yml" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 31 July 2024 21:24:42 -0400 (0:00:00.471) 0:00:56.343 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/wordpress:4.8-apache", "quay.io/linux-system-roles/envoyproxy:v1.25.0" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube", "__podman_volumes": [ "/tmp/httpd3", "/tmp/httpd3-create" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 31 July 2024 21:24:43 -0400 (0:00:00.118) 0:00:56.462 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 31 July 2024 21:24:43 -0400 (0:00:00.035) 0:00:56.497 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 31 July 2024 21:24:43 -0400 (0:00:00.027) 0:00:56.524 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Wednesday 31 July 2024 21:24:43 -0400 (0:00:00.095) 0:00:56.620 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:24:43 -0400 (0:00:00.052) 0:00:56.672 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:24:43 -0400 (0:00:00.029) 0:00:56.701 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:24:43 -0400 (0:00:00.030) 0:00:56.731 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Wednesday 31 July 2024 21:24:43 -0400 (0:00:00.027) 0:00:56.759 ******** [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => (item=/tmp/httpd3) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/httpd3", "mode": "0755", "owner": "root", "path": "/tmp/httpd3", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } changed: [managed_node1] => (item=/tmp/httpd3-create) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/tmp/httpd3-create", "mode": "0755", "owner": "root", "path": "/tmp/httpd3-create", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Wednesday 31 July 2024 21:24:44 -0400 (0:00:00.699) 0:00:57.458 ******** changed: [managed_node1] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [managed_node1] => (item=None) => { "attempts": 1, "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Wednesday 31 July 2024 21:25:01 -0400 (0:00:17.342) 0:01:14.801 ******** ok: [managed_node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 160, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:48 Wednesday 31 July 2024 21:25:01 -0400 (0:00:00.363) 0:01:15.165 ******** changed: [managed_node1] => { "changed": true, "checksum": "7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7", "dest": "/etc/containers/systemd/quadlet-demo.kube", "gid": 0, "group": "root", "md5sum": "da53c88f92b68b0487aa209f795b6bb3", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 456, "src": "/root/.ansible/tmp/ansible-tmp-1722475501.7691553-29033-234388092068636/source", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:58 Wednesday 31 July 2024 21:25:02 -0400 (0:00:00.671) 0:01:15.837 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:70 Wednesday 31 July 2024 21:25:02 -0400 (0:00:00.031) 0:01:15.868 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:82 Wednesday 31 July 2024 21:25:02 -0400 (0:00:00.031) 0:01:15.900 ******** ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 Wednesday 31 July 2024 21:25:03 -0400 (0:00:00.624) 0:01:16.524 ******** changed: [managed_node1] => { "changed": true, "name": "quadlet-demo.service", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "quadlet-demo-mysql.service systemd-journald.socket sysinit.target -.mount system.slice basic.target quadlet-demo-network.service", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo.service", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target quadlet-demo-network.service quadlet-demo-mysql.service -.mount system.slice", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo.kube", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:125 Wednesday 31 July 2024 21:25:04 -0400 (0:00:01.819) 0:01:18.343 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:143 Wednesday 31 July 2024 21:25:04 -0400 (0:00:00.033) 0:01:18.377 ******** TASK [fedora.linux_system_roles.podman : Handle credential files - absent] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:149 Wednesday 31 July 2024 21:25:04 -0400 (0:00:00.025) 0:01:18.402 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:158 Wednesday 31 July 2024 21:25:04 -0400 (0:00:00.028) 0:01:18.431 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [Check quadlet files] ***************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:84 Wednesday 31 July 2024 21:25:05 -0400 (0:00:00.025) 0:01:18.457 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "ls", "-alrtF", "/etc/containers/systemd" ], "delta": "0:00:00.003881", "end": "2024-07-31 21:25:05.360884", "rc": 0, "start": "2024-07-31 21:25:05.357003" } STDOUT: total 24 drwxr-xr-x. 9 root root 187 Jul 31 21:18 ../ -rw-r--r--. 1 root root 74 Jul 31 21:24 quadlet-demo.network -rw-r--r--. 1 root root 9 Jul 31 21:24 quadlet-demo-mysql.volume -rw-r--r--. 1 root root 363 Jul 31 21:24 quadlet-demo-mysql.container -rw-r--r--. 1 root root 2102 Jul 31 21:24 envoy-proxy-configmap.yml -rw-r--r--. 1 root root 1605 Jul 31 21:24 quadlet-demo.yml -rw-r--r--. 1 root root 456 Jul 31 21:25 quadlet-demo.kube drwxr-xr-x. 2 root root 185 Jul 31 21:25 ./ TASK [Check containers] ******************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:88 Wednesday 31 July 2024 21:25:05 -0400 (0:00:00.449) 0:01:18.906 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "-a" ], "delta": "0:00:00.049476", "end": "2024-07-31 21:25:05.859478", "failed_when_result": false, "rc": 0, "start": "2024-07-31 21:25:05.810002" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES ea182df3f7ef quay.io/linux-system-roles/mysql:5.6 mysqld 33 seconds ago Up 33 seconds (healthy) quadlet-demo-mysql c892d334f8f6 localhost/podman-pause:4.9.4-dev-1708535009 1 second ago Up 1 second a96f3a51b8d1-service cd265236ca47 localhost/podman-pause:4.9.4-dev-1708535009 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp 693c30debda5-infra 523c4ea94b7b quay.io/linux-system-roles/wordpress:4.8-apache apache2-foregroun... 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-wordpress 83d50c826b2d quay.io/linux-system-roles/envoyproxy:v1.25.0 envoy -c /etc/env... 1 second ago Up 1 second 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-envoy TASK [Check volumes] *********************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:93 Wednesday 31 July 2024 21:25:05 -0400 (0:00:00.471) 0:01:19.378 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls" ], "delta": "0:00:00.048364", "end": "2024-07-31 21:25:06.331757", "failed_when_result": false, "rc": 0, "start": "2024-07-31 21:25:06.283393" } STDOUT: DRIVER VOLUME NAME local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [Check pods] ************************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:98 Wednesday 31 July 2024 21:25:06 -0400 (0:00:00.462) 0:01:19.841 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "pod", "ps", "--ctr-ids", "--ctr-names", "--ctr-status" ], "delta": "0:00:00.037436", "end": "2024-07-31 21:25:06.731553", "failed_when_result": false, "rc": 0, "start": "2024-07-31 21:25:06.694117" } STDOUT: POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS 693c30debda5 quadlet-demo Running 2 seconds ago cd265236ca47 cd265236ca47,523c4ea94b7b,83d50c826b2d 693c30debda5-infra,quadlet-demo-wordpress,quadlet-demo-envoy running,running,running TASK [Check systemd] *********************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:103 Wednesday 31 July 2024 21:25:06 -0400 (0:00:00.397) 0:01:20.238 ******** ok: [managed_node1] => { "changed": false, "cmd": "set -euo pipefail; systemctl list-units | grep quadlet", "delta": "0:00:00.011103", "end": "2024-07-31 21:25:07.102659", "failed_when_result": false, "rc": 0, "start": "2024-07-31 21:25:07.091556" } STDOUT: quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service quadlet-demo-mysql.service loaded active running quadlet-demo-mysql.service quadlet-demo-network.service loaded active exited quadlet-demo-network.service quadlet-demo.service loaded active running quadlet-demo.service TASK [Check web] *************************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:109 Wednesday 31 July 2024 21:25:07 -0400 (0:00:00.369) 0:01:20.608 ******** changed: [managed_node1] => { "attempts": 1, "changed": true, "checksum_dest": null, "checksum_src": "9dbc8c154daaed74c52aa781ab9994c0246a012f", "dest": "/run/out", "elapsed": 0, "gid": 0, "group": "root", "md5sum": "f59bccb13786ba6aa549360a06d6b1b4", "mode": "0600", "owner": "root", "secontext": "system_u:object_r:var_run_t:s0", "size": 11666, "src": "/root/.ansible/tmp/ansible-tmp-1722475507.2050025-29124-224205375420076/tmpeglqj2xr", "state": "file", "status_code": 200, "uid": 0, "url": "https://localhost:8000" } MSG: OK (unknown bytes) TASK [Show web] **************************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:120 Wednesday 31 July 2024 21:25:08 -0400 (0:00:01.129) 0:01:21.737 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "cat", "/run/out" ], "delta": "0:00:00.006042", "end": "2024-07-31 21:25:08.591162", "rc": 0, "start": "2024-07-31 21:25:08.585120" } STDOUT: WordPress › Installation

TASK [Error] ******************************************************************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:125 Wednesday 31 July 2024 21:25:08 -0400 (0:00:00.356) 0:01:22.094 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check] ******************************************************************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:136 Wednesday 31 July 2024 21:25:08 -0400 (0:00:00.057) 0:01:22.152 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "-a" ], "delta": "0:00:00.035205", "end": "2024-07-31 21:25:09.058125", "rc": 0, "start": "2024-07-31 21:25:09.022920" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES ea182df3f7ef quay.io/linux-system-roles/mysql:5.6 mysqld 36 seconds ago Up 37 seconds (healthy) quadlet-demo-mysql c892d334f8f6 localhost/podman-pause:4.9.4-dev-1708535009 5 seconds ago Up 5 seconds a96f3a51b8d1-service cd265236ca47 localhost/podman-pause:4.9.4-dev-1708535009 5 seconds ago Up 5 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp 693c30debda5-infra 523c4ea94b7b quay.io/linux-system-roles/wordpress:4.8-apache apache2-foregroun... 5 seconds ago Up 5 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-wordpress 83d50c826b2d quay.io/linux-system-roles/envoyproxy:v1.25.0 envoy -c /etc/env... 5 seconds ago Up 5 seconds 0.0.0.0:8000->8080/tcp, 0.0.0.0:9000->9901/tcp quadlet-demo-envoy TASK [Check pods] ************************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:140 Wednesday 31 July 2024 21:25:09 -0400 (0:00:00.409) 0:01:22.561 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "pod", "ps", "--ctr-ids", "--ctr-names", "--ctr-status" ], "delta": "0:00:00.038344", "end": "2024-07-31 21:25:09.443303", "failed_when_result": false, "rc": 0, "start": "2024-07-31 21:25:09.404959" } STDOUT: POD ID NAME STATUS CREATED INFRA ID IDS NAMES STATUS 693c30debda5 quadlet-demo Running 5 seconds ago cd265236ca47 cd265236ca47,523c4ea94b7b,83d50c826b2d 693c30debda5-infra,quadlet-demo-wordpress,quadlet-demo-envoy running,running,running TASK [Check systemd] *********************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:145 Wednesday 31 July 2024 21:25:09 -0400 (0:00:00.384) 0:01:22.946 ******** ok: [managed_node1] => { "changed": false, "cmd": "set -euo pipefail; systemctl list-units --all | grep quadlet", "delta": "0:00:00.009969", "end": "2024-07-31 21:25:09.796389", "failed_when_result": false, "rc": 0, "start": "2024-07-31 21:25:09.786420" } STDOUT: quadlet-demo-mysql-volume.service loaded active exited quadlet-demo-mysql-volume.service quadlet-demo-mysql.service loaded active running quadlet-demo-mysql.service quadlet-demo-network.service loaded active exited quadlet-demo-network.service quadlet-demo.service loaded active running quadlet-demo.service TASK [LS] ********************************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:153 Wednesday 31 July 2024 21:25:09 -0400 (0:00:00.352) 0:01:23.299 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "ls", "-alrtF", "/etc/systemd/system" ], "delta": "0:00:01.004780", "end": "2024-07-31 21:25:11.148711", "failed_when_result": false, "rc": 0, "start": "2024-07-31 21:25:10.143931" } STDOUT: total 12 lrwxrwxrwx. 1 root root 9 May 11 2019 systemd-timedated.service -> /dev/null drwxr-xr-x. 4 root root 169 May 29 03:45 ../ lrwxrwxrwx. 1 root root 39 May 29 03:45 syslog.service -> /usr/lib/systemd/system/rsyslog.service drwxr-xr-x. 2 root root 32 May 29 03:45 getty.target.wants/ lrwxrwxrwx. 1 root root 37 May 29 03:45 ctrl-alt-del.target -> /usr/lib/systemd/system/reboot.target lrwxrwxrwx. 1 root root 57 May 29 03:45 dbus-org.freedesktop.nm-dispatcher.service -> /usr/lib/systemd/system/NetworkManager-dispatcher.service drwxr-xr-x. 2 root root 48 May 29 03:45 network-online.target.wants/ lrwxrwxrwx. 1 root root 41 May 29 03:45 dbus-org.freedesktop.timedate1.service -> /usr/lib/systemd/system/timedatex.service drwxr-xr-x. 2 root root 61 May 29 03:45 timers.target.wants/ drwxr-xr-x. 2 root root 31 May 29 03:45 basic.target.wants/ drwxr-xr-x. 2 root root 38 May 29 03:45 dev-virtio\x2dports-org.qemu.guest_agent.0.device.wants/ lrwxrwxrwx. 1 root root 41 May 29 03:47 default.target -> /usr/lib/systemd/system/multi-user.target drwxr-xr-x. 2 root root 51 May 29 03:55 sockets.target.wants/ drwxr-xr-x. 2 root root 31 May 29 03:55 remote-fs.target.wants/ drwxr-xr-x. 2 root root 59 May 29 03:55 sshd-keygen@.service.d/ drwxr-xr-x. 2 root root 119 May 29 03:55 cloud-init.target.wants/ drwxr-xr-x. 2 root root 181 May 29 03:56 sysinit.target.wants/ drwxr-xr-x. 2 root root 4096 Jul 31 21:13 multi-user.target.wants/ lrwxrwxrwx. 1 root root 41 Jul 31 21:13 dbus-org.fedoraproject.FirewallD1.service -> /usr/lib/systemd/system/firewalld.service -rw-r--r--. 1 root root 99 Jul 31 21:19 user-0.slice drwxr-xr-x. 2 root root 27 Jul 31 21:19 user@.service.d/ drwxr-xr-x. 2 root root 27 Jul 31 21:19 user-.slice.d/ drwxr-xr-x. 15 root root 4096 Jul 31 21:19 ./ TASK [Cleanup] ***************************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:160 Wednesday 31 July 2024 21:25:11 -0400 (0:00:01.357) 0:01:24.657 ******** TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Wednesday 31 July 2024 21:25:11 -0400 (0:00:00.062) 0:01:24.719 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Wednesday 31 July 2024 21:25:11 -0400 (0:00:00.088) 0:01:24.807 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Wednesday 31 July 2024 21:25:11 -0400 (0:00:00.037) 0:01:24.845 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Wednesday 31 July 2024 21:25:11 -0400 (0:00:00.030) 0:01:24.875 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:20 Wednesday 31 July 2024 21:25:11 -0400 (0:00:00.027) 0:01:24.902 ******** ok: [managed_node1] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed_node1] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node1] => (item=CentOS_8.yml) => { "ansible_facts": { "__podman_packages": [ "crun", "podman", "podman-plugins", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml" } ok: [managed_node1] => (item=CentOS_8.yml) => { "ansible_facts": { "__podman_packages": [ "crun", "podman", "podman-plugins", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/vars/CentOS_8.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_8.yml" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Wednesday 31 July 2024 21:25:11 -0400 (0:00:00.079) 0:01:24.981 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Wednesday 31 July 2024 21:25:12 -0400 (0:00:01.411) 0:01:26.393 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Wednesday 31 July 2024 21:25:12 -0400 (0:00:00.028) 0:01:26.421 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:22 Wednesday 31 July 2024 21:25:13 -0400 (0:00:00.038) 0:01:26.460 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.027823", "end": "2024-07-31 21:25:13.325464", "rc": 0, "start": "2024-07-31 21:25:13.297641" } STDOUT: podman version 4.9.4-dev TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Wednesday 31 July 2024 21:25:13 -0400 (0:00:00.369) 0:01:26.829 ******** ok: [managed_node1] => { "ansible_facts": { "podman_version": "4.9.4-dev" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:32 Wednesday 31 July 2024 21:25:13 -0400 (0:00:00.032) 0:01:26.862 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:39 Wednesday 31 July 2024 21:25:13 -0400 (0:00:00.028) 0:01:26.891 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } META: end_host conditional evaluated to false, continuing execution for managed_node1 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Wednesday 31 July 2024 21:25:13 -0400 (0:00:00.057) 0:01:26.948 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:25:13 -0400 (0:00:00.062) 0:01:27.011 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:25:13 -0400 (0:00:00.039) 0:01:27.050 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:25:13 -0400 (0:00:00.035) 0:01:27.086 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:25:13 -0400 (0:00:00.041) 0:01:27.127 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.362) 0:01:27.490 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.038) 0:01:27.529 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.374) 0:01:27.904 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.031) 0:01:27.936 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.030) 0:01:27.966 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.061) 0:01:28.027 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.032) 0:01:28.060 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.028) 0:01:28.088 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.030) 0:01:28.119 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.027) 0:01:28.146 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:62 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.031) 0:01:28.178 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:71 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.041) 0:01:28.219 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.060) 0:01:28.280 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.029) 0:01:28.309 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:74 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.030) 0:01:28.340 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.060) 0:01:28.400 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Wednesday 31 July 2024 21:25:14 -0400 (0:00:00.029) 0:01:28.430 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:77 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.032) 0:01:28.462 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:5 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.061) 0:01:28.524 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:13 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.028) 0:01:28.553 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.063) 0:01:28.616 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:6 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.067) 0:01:28.684 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:14 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.030) 0:01:28.714 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:19 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.031) 0:01:28.745 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:25 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.029) 0:01:28.775 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:86 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.031) 0:01:28.806 ******** TASK [fedora.linux_system_roles.firewall : Setup firewalld] ******************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.089) 0:01:28.896 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed_node1 TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.056) 0:01:28.953 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if system is ostree] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.038) 0:01:28.991 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.030) 0:01:29.021 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.031) 0:01:29.053 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.029) 0:01:29.082 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Install firewalld] ****************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 Wednesday 31 July 2024 21:25:15 -0400 (0:00:00.032) 0:01:29.114 ******** ok: [managed_node1] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43 Wednesday 31 July 2024 21:25:18 -0400 (0:00:02.404) 0:01:31.518 ******** skipping: [managed_node1] => {} TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48 Wednesday 31 July 2024 21:25:18 -0400 (0:00:00.029) 0:01:31.548 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53 Wednesday 31 July 2024 21:25:18 -0400 (0:00:00.030) 0:01:31.579 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Collect service facts] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5 Wednesday 31 July 2024 21:25:18 -0400 (0:00:00.028) 0:01:31.607 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9 Wednesday 31 July 2024 21:25:18 -0400 (0:00:00.030) 0:01:31.637 ******** skipping: [managed_node1] => (item=nftables) => { "ansible_loop_var": "item", "changed": false, "item": "nftables", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=iptables) => { "ansible_loop_var": "item", "changed": false, "item": "iptables", "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item=ufw) => { "ansible_loop_var": "item", "changed": false, "item": "ufw", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 Wednesday 31 July 2024 21:25:18 -0400 (0:00:00.036) 0:01:31.674 ******** ok: [managed_node1] => { "changed": false, "name": "firewalld", "status": { "ActiveEnterTimestamp": "Wed 2024-07-31 21:19:43 EDT", "ActiveEnterTimestampMonotonic": "6479081", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "sysinit.target dbus.service basic.target system.slice polkit.service dbus.socket", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-31 21:19:42 EDT", "AssertTimestampMonotonic": "5431503", "Before": "shutdown.target network-pre.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-31 21:19:42 EDT", "ConditionTimestampMonotonic": "5431502", "ConfigurationDirectoryMode": "0755", "Conflicts": "ip6tables.service shutdown.target nftables.service ipset.service ebtables.service iptables.service", "ControlGroup": "/system.slice/firewalld.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "man:firewalld(1)", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "582", "ExecMainStartTimestamp": "Wed 2024-07-31 21:19:42 EDT", "ExecMainStartTimestampMonotonic": "5433279", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-31 21:19:42 EDT", "InactiveExitTimestampMonotonic": "5433306", "InvocationID": "a1cc5d84d60d46859a9019e5cb14bc27", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "582", "MemoryAccounting": "yes", "MemoryCurrent": "54190080", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Wed 2024-07-31 21:19:43 EDT", "StateChangeTimestampMonotonic": "6479081", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogTimestamp": "Wed 2024-07-31 21:19:43 EDT", "WatchdogTimestampMonotonic": "6479078", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 Wednesday 31 July 2024 21:25:18 -0400 (0:00:00.484) 0:01:32.159 ******** ok: [managed_node1] => { "changed": false, "enabled": true, "name": "firewalld", "state": "started", "status": { "ActiveEnterTimestamp": "Wed 2024-07-31 21:19:43 EDT", "ActiveEnterTimestampMonotonic": "6479081", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "sysinit.target dbus.service basic.target system.slice polkit.service dbus.socket", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-31 21:19:42 EDT", "AssertTimestampMonotonic": "5431503", "Before": "shutdown.target network-pre.target multi-user.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "org.fedoraproject.FirewallD1", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-31 21:19:42 EDT", "ConditionTimestampMonotonic": "5431502", "ConfigurationDirectoryMode": "0755", "Conflicts": "ip6tables.service shutdown.target nftables.service ipset.service ebtables.service iptables.service", "ControlGroup": "/system.slice/firewalld.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "firewalld - dynamic firewall daemon", "DevicePolicy": "auto", "Documentation": "man:firewalld(1)", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "582", "ExecMainStartTimestamp": "Wed 2024-07-31 21:19:42 EDT", "ExecMainStartTimestampMonotonic": "5433279", "ExecMainStatus": "0", "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/firewalld.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "firewalld.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-31 21:19:42 EDT", "InactiveExitTimestampMonotonic": "5433306", "InvocationID": "a1cc5d84d60d46859a9019e5cb14bc27", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "582", "MemoryAccounting": "yes", "MemoryCurrent": "54190080", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "firewalld.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "null", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "null", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Wed 2024-07-31 21:19:43 EDT", "StateChangeTimestampMonotonic": "6479081", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "2", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "Wants": "network-pre.target", "WatchdogTimestamp": "Wed 2024-07-31 21:19:43 EDT", "WatchdogTimestampMonotonic": "6479078", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34 Wednesday 31 July 2024 21:25:19 -0400 (0:00:00.486) 0:01:32.645 ******** ok: [managed_node1] => { "ansible_facts": { "__firewall_previous_replaced": false, "__firewall_python_cmd": "/usr/libexec/platform-python", "__firewall_report_changed": true }, "changed": false } TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43 Wednesday 31 July 2024 21:25:19 -0400 (0:00:00.035) 0:01:32.680 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55 Wednesday 31 July 2024 21:25:19 -0400 (0:00:00.029) 0:01:32.710 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Configure firewall] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 Wednesday 31 July 2024 21:25:19 -0400 (0:00:00.028) 0:01:32.739 ******** ok: [managed_node1] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" } } ok: [managed_node1] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "__firewall_changed": false, "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" } } TASK [fedora.linux_system_roles.firewall : Gather firewall config information] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120 Wednesday 31 July 2024 21:25:20 -0400 (0:00:01.074) 0:01:33.813 ******** skipping: [managed_node1] => (item={'port': '8000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "8000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } skipping: [managed_node1] => (item={'port': '9000/tcp', 'state': 'enabled'}) => { "ansible_loop_var": "item", "changed": false, "item": { "port": "9000/tcp", "state": "enabled" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.047) 0:01:33.861 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.033) 0:01:33.894 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.029) 0:01:33.924 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.031) 0:01:33.955 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Calculate what has changed] ********* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.027) 0:01:33.982 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.firewall : Show diffs] ************************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.030) 0:01:34.013 ******** skipping: [managed_node1] => {} TASK [Manage selinux for specified ports] ************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:93 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.026) 0:01:34.040 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:100 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.059) 0:01:34.100 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:104 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.030) 0:01:34.131 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:113 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.029) 0:01:34.160 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:122 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.031) 0:01:34.191 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.092) 0:01:34.283 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.035) 0:01:34.318 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.039) 0:01:34.358 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.050) 0:01:34.409 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:25:20 -0400 (0:00:00.029) 0:01:34.439 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:25:21 -0400 (0:00:00.033) 0:01:34.472 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Wednesday 31 July 2024 21:25:21 -0400 (0:00:00.028) 0:01:34.501 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Wednesday 31 July 2024 21:25:21 -0400 (0:00:00.033) 0:01:34.534 ******** [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Wednesday 31 July 2024 21:25:21 -0400 (0:00:00.400) 0:01:34.934 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Wednesday 31 July 2024 21:25:21 -0400 (0:00:00.033) 0:01:34.967 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Wednesday 31 July 2024 21:25:21 -0400 (0:00:00.073) 0:01:35.041 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:25:21 -0400 (0:00:00.049) 0:01:35.091 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:25:21 -0400 (0:00:00.030) 0:01:35.122 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:25:21 -0400 (0:00:00.029) 0:01:35.151 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Wednesday 31 July 2024 21:25:21 -0400 (0:00:00.030) 0:01:35.182 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Wednesday 31 July 2024 21:25:21 -0400 (0:00:00.028) 0:01:35.210 ******** [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Set variables part 1] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:3 Wednesday 31 July 2024 21:25:22 -0400 (0:00:00.395) 0:01:35.606 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set variables part 2] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:7 Wednesday 31 July 2024 21:25:22 -0400 (0:00:00.031) 0:01:35.637 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_rootless": false, "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:13 Wednesday 31 July 2024 21:25:22 -0400 (0:00:00.038) 0:01:35.675 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:25:22 -0400 (0:00:00.049) 0:01:35.724 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:25:22 -0400 (0:00:00.029) 0:01:35.753 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:25:22 -0400 (0:00:00.030) 0:01:35.784 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:18 Wednesday 31 July 2024 21:25:22 -0400 (0:00:00.028) 0:01:35.812 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Manage each secret] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_secret.yml:34 Wednesday 31 July 2024 21:25:22 -0400 (0:00:00.028) 0:01:35.841 ******** [WARNING]: Using a variable for a task's 'args' is unsafe in some situations (see https://docs.ansible.com/ansible/devel/reference_appendices/faq.html#argsplat- unsafe) changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:129 Wednesday 31 July 2024 21:25:22 -0400 (0:00:00.418) 0:01:36.260 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:136 Wednesday 31 July 2024 21:25:22 -0400 (0:00:00.027) 0:01:36.288 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 31 July 2024 21:25:22 -0400 (0:00:00.137) 0:01:36.426 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.kube", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Unit]\nRequires=quadlet-demo-mysql.service\nAfter=quadlet-demo-mysql.service\n\n[Kube]\n# Point to the yaml file in the same directory\nYaml=quadlet-demo.yml\n# Use the quadlet-demo network\nNetwork=quadlet-demo.network\n# Publish the envoy proxy data port\nPublishPort=8000:8080\n# Publish the envoy proxy admin port\nPublishPort=9000:9901\n# Use the envoy proxy config map in the same directory\nConfigMap=envoy-proxy-configmap.yml", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 31 July 2024 21:25:23 -0400 (0:00:00.039) 0:01:36.465 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 31 July 2024 21:25:23 -0400 (0:00:00.037) 0:01:36.502 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 31 July 2024 21:25:23 -0400 (0:00:00.031) 0:01:36.534 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "kube", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 31 July 2024 21:25:23 -0400 (0:00:00.042) 0:01:36.576 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:25:23 -0400 (0:00:00.056) 0:01:36.632 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:25:23 -0400 (0:00:00.038) 0:01:36.671 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:25:23 -0400 (0:00:00.035) 0:01:36.707 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:25:23 -0400 (0:00:00.039) 0:01:36.746 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:25:23 -0400 (0:00:00.356) 0:01:37.102 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:25:23 -0400 (0:00:00.038) 0:01:37.141 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.352) 0:01:37.494 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.029) 0:01:37.524 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.030) 0:01:37.554 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.027) 0:01:37.582 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.031) 0:01:37.613 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.027) 0:01:37.641 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.030) 0:01:37.672 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.027) 0:01:37.699 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.030) 0:01:37.730 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": [ "quadlet-demo.yml" ], "__podman_service_name": "quadlet-demo.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.053) 0:01:37.783 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.029) 0:01:37.812 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.028) 0:01:37.841 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.kube", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.069) 0:01:37.910 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.032) 0:01:37.943 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.073) 0:01:38.017 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Wednesday 31 July 2024 21:25:24 -0400 (0:00:00.028) 0:01:38.045 ******** changed: [managed_node1] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo.service", "state": "stopped", "status": { "ActiveEnterTimestamp": "Wed 2024-07-31 21:25:04 EDT", "ActiveEnterTimestampMonotonic": "326631451", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "-.mount system.slice quadlet-demo-network.service systemd-journald.socket quadlet-demo-mysql.service basic.target sysinit.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-31 21:25:03 EDT", "AssertTimestampMonotonic": "325348075", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-31 21:25:03 EDT", "ConditionTimestampMonotonic": "325348073", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroup": "/system.slice/quadlet-demo.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo.service", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "60053", "ExecMainStartTimestamp": "Wed 2024-07-31 21:25:04 EDT", "ExecMainStartTimestampMonotonic": "326631429", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube play --replace --service-container=true --network=systemd-quadlet-demo --configmap /etc/containers/systemd/envoy-proxy-configmap.yml --publish 8000:8080 --publish 9000:9901 /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[Wed 2024-07-31 21:25:03 EDT] ; stop_time=[n/a] ; pid=59942 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman kube down /etc/containers/systemd/quadlet-demo.yml ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-31 21:25:03 EDT", "InactiveExitTimestampMonotonic": "325349172", "InvocationID": "7ab69916603c444fa6aa5adce63cab0b", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "60053", "MemoryAccounting": "yes", "MemoryCurrent": "6164480", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "quadlet-demo-network.service sysinit.target quadlet-demo-mysql.service -.mount system.slice", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo.kube", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Wed 2024-07-31 21:25:04 EDT", "StateChangeTimestampMonotonic": "326631451", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "4", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestamp": "Wed 2024-07-31 21:25:04 EDT", "WatchdogTimestampMonotonic": "326631448", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Wednesday 31 July 2024 21:25:25 -0400 (0:00:01.192) 0:01:39.237 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722475502.922543, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7", "ctime": 1722475502.3245418, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 7135334, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1722475502.0645413, "nlink": 1, "path": "/etc/containers/systemd/quadlet-demo.kube", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 456, "uid": 0, "version": "581181250", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Wednesday 31 July 2024 21:25:26 -0400 (0:00:00.356) 0:01:39.594 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Wednesday 31 July 2024 21:25:26 -0400 (0:00:00.055) 0:01:39.649 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Wednesday 31 July 2024 21:25:26 -0400 (0:00:00.343) 0:01:39.993 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Wednesday 31 July 2024 21:25:26 -0400 (0:00:00.046) 0:01:40.039 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Wednesday 31 July 2024 21:25:26 -0400 (0:00:00.028) 0:01:40.068 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Wednesday 31 July 2024 21:25:26 -0400 (0:00:00.030) 0:01:40.098 ******** changed: [managed_node1] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.kube", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Wednesday 31 July 2024 21:25:27 -0400 (0:00:00.357) 0:01:40.456 ******** ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Wednesday 31 July 2024 21:25:27 -0400 (0:00:00.589) 0:01:41.045 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95 Wednesday 31 July 2024 21:25:27 -0400 (0:00:00.035) 0:01:41.081 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Wednesday 31 July 2024 21:25:27 -0400 (0:00:00.041) 0:01:41.123 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Wednesday 31 July 2024 21:25:27 -0400 (0:00:00.031) 0:01:41.154 ******** changed: [managed_node1] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.606494", "end": "2024-07-31 21:25:28.612539", "rc": 0, "start": "2024-07-31 21:25:28.006045" } STDOUT: fcf3e41b8864a14d75a6d0627d3d02154e28a153aa57e8baa392cd744ffa0d0b 5af2585e22ed1562885d9407efab74010090427be79048c2cd6a226517cc1e1d 7b9a6d05497a937dc611d9e3e13384aebc3d85eb9f404d961082b6bd4e0051d3 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127 Wednesday 31 July 2024 21:25:28 -0400 (0:00:00.962) 0:01:42.117 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:25:28 -0400 (0:00:00.056) 0:01:42.174 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:25:28 -0400 (0:00:00.031) 0:01:42.206 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:25:28 -0400 (0:00:00.029) 0:01:42.235 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137 Wednesday 31 July 2024 21:25:28 -0400 (0:00:00.031) 0:01:42.267 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.031969", "end": "2024-07-31 21:25:29.154871", "rc": 0, "start": "2024-07-31 21:25:29.122902" } STDOUT: quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146 Wednesday 31 July 2024 21:25:29 -0400 (0:00:00.396) 0:01:42.664 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.030470", "end": "2024-07-31 21:25:29.546992", "rc": 0, "start": "2024-07-31 21:25:29.516522" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155 Wednesday 31 July 2024 21:25:29 -0400 (0:00:00.428) 0:01:43.093 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.034881", "end": "2024-07-31 21:25:29.976276", "rc": 0, "start": "2024-07-31 21:25:29.941395" } STDOUT: ea182df3f7ef quay.io/linux-system-roles/mysql:5.6 mysqld 57 seconds ago Up 57 seconds (healthy) quadlet-demo-mysql TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164 Wednesday 31 July 2024 21:25:30 -0400 (0:00:00.389) 0:01:43.483 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.054018", "end": "2024-07-31 21:25:30.392582", "rc": 0, "start": "2024-07-31 21:25:30.338564" } STDOUT: podman podman-default-kube-network systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173 Wednesday 31 July 2024 21:25:30 -0400 (0:00:00.414) 0:01:43.897 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 Wednesday 31 July 2024 21:25:30 -0400 (0:00:00.385) 0:01:44.283 ******** ok: [managed_node1] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cni-dhcp.service": { "name": "cni-dhcp.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnsmasq.service": { "name": "dnsmasq.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ea182df3f7eff7c985466c...d656e421db49137275cc20e2f48b59c608623.service": { "name": "ea182df3f7eff7c985466c...d656e421db49137275cc20e2f48b59c608623.service", "source": "systemd", "state": "inactive", "status": "transient" }, "ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service": { "name": "ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "running", "status": "generated" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 31 July 2024 21:25:32 -0400 (0:00:01.676) 0:01:45.959 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 31 July 2024 21:25:32 -0400 (0:00:00.062) 0:01:46.021 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: PersistentVolumeClaim\nmetadata:\n name: wp-pv-claim\n labels:\n app: wordpress\nspec:\n accessModes:\n - ReadWriteOnce\n resources:\n requests:\n storage: 20Gi\n---\napiVersion: v1\nkind: Pod\nmetadata:\n name: quadlet-demo\nspec:\n containers:\n - name: wordpress\n image: quay.io/linux-system-roles/wordpress:4.8-apache\n env:\n - name: WORDPRESS_DB_HOST\n value: quadlet-demo-mysql\n - name: WORDPRESS_DB_PASSWORD\n valueFrom:\n secretKeyRef:\n name: mysql-root-password-kube\n key: password\n volumeMounts:\n - name: wordpress-persistent-storage\n mountPath: /var/www/html\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n - name: envoy\n image: quay.io/linux-system-roles/envoyproxy:v1.25.0\n volumeMounts:\n - name: config-volume\n mountPath: /etc/envoy\n - name: certificates\n mountPath: /etc/envoy-certificates\n env:\n - name: ENVOY_UID\n value: \"0\"\n resources:\n requests:\n memory: \"64Mi\"\n cpu: \"250m\"\n limits:\n memory: \"128Mi\"\n cpu: \"500m\"\n volumes:\n - name: config-volume\n configMap:\n name: envoy-proxy-config\n - name: certificates\n secret:\n secretName: envoy-certificates\n - name: wordpress-persistent-storage\n persistentVolumeClaim:\n claimName: wp-pv-claim\n - name: www # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3\n - name: create # not used - for testing hostpath\n hostPath:\n path: /tmp/httpd3-create\n", "__podman_quadlet_template_src": "quadlet-demo.yml.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 31 July 2024 21:25:32 -0400 (0:00:00.107) 0:01:46.128 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 31 July 2024 21:25:32 -0400 (0:00:00.040) 0:01:46.169 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 31 July 2024 21:25:32 -0400 (0:00:00.033) 0:01:46.203 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 31 July 2024 21:25:32 -0400 (0:00:00.045) 0:01:46.248 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:25:32 -0400 (0:00:00.062) 0:01:46.311 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:25:32 -0400 (0:00:00.040) 0:01:46.351 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:25:32 -0400 (0:00:00.038) 0:01:46.390 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:25:32 -0400 (0:00:00.042) 0:01:46.433 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:25:33 -0400 (0:00:00.359) 0:01:46.792 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:25:33 -0400 (0:00:00.041) 0:01:46.833 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:25:33 -0400 (0:00:00.348) 0:01:47.181 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:25:33 -0400 (0:00:00.032) 0:01:47.214 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:25:33 -0400 (0:00:00.068) 0:01:47.283 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:25:33 -0400 (0:00:00.032) 0:01:47.316 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:25:33 -0400 (0:00:00.033) 0:01:47.350 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:25:33 -0400 (0:00:00.031) 0:01:47.381 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:25:33 -0400 (0:00:00.033) 0:01:47.415 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:25:34 -0400 (0:00:00.032) 0:01:47.447 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 31 July 2024 21:25:34 -0400 (0:00:00.030) 0:01:47.478 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 31 July 2024 21:25:34 -0400 (0:00:00.057) 0:01:47.536 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 31 July 2024 21:25:34 -0400 (0:00:00.033) 0:01:47.570 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 31 July 2024 21:25:34 -0400 (0:00:00.030) 0:01:47.601 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 31 July 2024 21:25:34 -0400 (0:00:00.070) 0:01:47.671 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 31 July 2024 21:25:34 -0400 (0:00:00.038) 0:01:47.709 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Wednesday 31 July 2024 21:25:34 -0400 (0:00:00.071) 0:01:47.781 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Wednesday 31 July 2024 21:25:34 -0400 (0:00:00.033) 0:01:47.815 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Wednesday 31 July 2024 21:25:34 -0400 (0:00:00.031) 0:01:47.846 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722475482.8144996, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "998dccde0483b1654327a46ddd89cbaa47650370", "ctime": 1722475480.198494, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 513802631, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1722475479.8304932, "nlink": 1, "path": "/etc/containers/systemd/quadlet-demo.yml", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1605, "uid": 0, "version": "1928184577", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Wednesday 31 July 2024 21:25:34 -0400 (0:00:00.384) 0:01:48.230 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Wednesday 31 July 2024 21:25:34 -0400 (0:00:00.061) 0:01:48.292 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Wednesday 31 July 2024 21:25:35 -0400 (0:00:00.348) 0:01:48.640 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Wednesday 31 July 2024 21:25:35 -0400 (0:00:00.033) 0:01:48.674 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Wednesday 31 July 2024 21:25:35 -0400 (0:00:00.037) 0:01:48.712 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Wednesday 31 July 2024 21:25:35 -0400 (0:00:00.032) 0:01:48.745 ******** changed: [managed_node1] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Wednesday 31 July 2024 21:25:35 -0400 (0:00:00.351) 0:01:49.097 ******** ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Wednesday 31 July 2024 21:25:36 -0400 (0:00:00.594) 0:01:49.691 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95 Wednesday 31 July 2024 21:25:36 -0400 (0:00:00.037) 0:01:49.729 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Wednesday 31 July 2024 21:25:36 -0400 (0:00:00.085) 0:01:49.815 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Wednesday 31 July 2024 21:25:36 -0400 (0:00:00.031) 0:01:49.846 ******** changed: [managed_node1] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.034539", "end": "2024-07-31 21:25:36.744845", "rc": 0, "start": "2024-07-31 21:25:36.710306" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127 Wednesday 31 July 2024 21:25:36 -0400 (0:00:00.407) 0:01:50.254 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:25:36 -0400 (0:00:00.060) 0:01:50.314 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:25:36 -0400 (0:00:00.033) 0:01:50.347 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:25:36 -0400 (0:00:00.032) 0:01:50.380 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137 Wednesday 31 July 2024 21:25:36 -0400 (0:00:00.034) 0:01:50.414 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.032121", "end": "2024-07-31 21:25:37.298355", "rc": 0, "start": "2024-07-31 21:25:37.266234" } STDOUT: quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146 Wednesday 31 July 2024 21:25:37 -0400 (0:00:00.392) 0:01:50.807 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.031191", "end": "2024-07-31 21:25:37.686034", "rc": 0, "start": "2024-07-31 21:25:37.654843" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155 Wednesday 31 July 2024 21:25:37 -0400 (0:00:00.387) 0:01:51.194 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.031917", "end": "2024-07-31 21:25:38.076346", "rc": 0, "start": "2024-07-31 21:25:38.044429" } STDOUT: ea182df3f7ef quay.io/linux-system-roles/mysql:5.6 mysqld About a minute ago Up About a minute (healthy) quadlet-demo-mysql TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164 Wednesday 31 July 2024 21:25:38 -0400 (0:00:00.390) 0:01:51.585 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.053254", "end": "2024-07-31 21:25:38.494694", "rc": 0, "start": "2024-07-31 21:25:38.441440" } STDOUT: podman podman-default-kube-network systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173 Wednesday 31 July 2024 21:25:38 -0400 (0:00:00.419) 0:01:52.004 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 Wednesday 31 July 2024 21:25:38 -0400 (0:00:00.385) 0:01:52.389 ******** ok: [managed_node1] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cni-dhcp.service": { "name": "cni-dhcp.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnsmasq.service": { "name": "dnsmasq.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ea182df3f7eff7c985466c...d656e421db49137275cc20e2f48b59c608623.service": { "name": "ea182df3f7eff7c985466c...d656e421db49137275cc20e2f48b59c608623.service", "source": "systemd", "state": "inactive", "status": "transient" }, "ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service": { "name": "ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "running", "status": "generated" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 31 July 2024 21:25:40 -0400 (0:00:01.537) 0:01:53.927 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 31 July 2024 21:25:40 -0400 (0:00:00.033) 0:01:53.960 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "envoy-proxy-configmap.yml", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "---\napiVersion: v1\nkind: ConfigMap\nmetadata:\n name: envoy-proxy-config\ndata:\n envoy.yaml: |\n admin:\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 9901\n\n static_resources:\n listeners:\n - name: listener_0\n address:\n socket_address:\n address: 0.0.0.0\n port_value: 8080\n filter_chains:\n - filters:\n - name: envoy.filters.network.http_connection_manager\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager\n stat_prefix: ingress_http\n codec_type: AUTO\n route_config:\n name: local_route\n virtual_hosts:\n - name: local_service\n domains: [\"*\"]\n routes:\n - match:\n prefix: \"/\"\n route:\n cluster: backend\n http_filters:\n - name: envoy.filters.http.router\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.filters.http.router.v3.Router\n transport_socket:\n name: envoy.transport_sockets.tls\n typed_config:\n \"@type\": type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext\n common_tls_context:\n tls_certificates:\n - certificate_chain:\n filename: /etc/envoy-certificates/certificate.pem\n private_key:\n filename: /etc/envoy-certificates/certificate.key\n clusters:\n - name: backend\n connect_timeout: 5s\n type: STATIC\n dns_refresh_rate: 1800s\n lb_policy: ROUND_ROBIN\n load_assignment:\n cluster_name: backend\n endpoints:\n - lb_endpoints:\n - endpoint:\n address:\n socket_address:\n address: 127.0.0.1\n port_value: 80", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 31 July 2024 21:25:40 -0400 (0:00:00.046) 0:01:54.007 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 31 July 2024 21:25:40 -0400 (0:00:00.039) 0:01:54.046 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 31 July 2024 21:25:40 -0400 (0:00:00.035) 0:01:54.081 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "envoy-proxy-configmap", "__podman_quadlet_type": "yml", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 31 July 2024 21:25:40 -0400 (0:00:00.081) 0:01:54.163 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:25:40 -0400 (0:00:00.066) 0:01:54.230 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:25:40 -0400 (0:00:00.038) 0:01:54.269 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:25:40 -0400 (0:00:00.040) 0:01:54.309 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:25:40 -0400 (0:00:00.045) 0:01:54.355 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:25:41 -0400 (0:00:00.361) 0:01:54.716 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:25:41 -0400 (0:00:00.041) 0:01:54.758 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:25:41 -0400 (0:00:00.356) 0:01:55.114 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:25:41 -0400 (0:00:00.032) 0:01:55.147 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:25:41 -0400 (0:00:00.032) 0:01:55.179 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:25:41 -0400 (0:00:00.031) 0:01:55.211 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:25:41 -0400 (0:00:00.034) 0:01:55.245 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:25:41 -0400 (0:00:00.075) 0:01:55.320 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:25:41 -0400 (0:00:00.034) 0:01:55.355 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:25:41 -0400 (0:00:00.032) 0:01:55.388 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 31 July 2024 21:25:41 -0400 (0:00:00.034) 0:01:55.422 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 31 July 2024 21:25:42 -0400 (0:00:00.055) 0:01:55.478 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 31 July 2024 21:25:42 -0400 (0:00:00.033) 0:01:55.511 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 31 July 2024 21:25:42 -0400 (0:00:00.032) 0:01:55.544 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/envoy-proxy-configmap.yml", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 31 July 2024 21:25:42 -0400 (0:00:00.073) 0:01:55.618 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 31 July 2024 21:25:42 -0400 (0:00:00.038) 0:01:55.656 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Wednesday 31 July 2024 21:25:42 -0400 (0:00:00.075) 0:01:55.732 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Wednesday 31 July 2024 21:25:42 -0400 (0:00:00.031) 0:01:55.763 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Wednesday 31 July 2024 21:25:42 -0400 (0:00:00.032) 0:01:55.796 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722475503.9145453, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "d681c7d56f912150d041873e880818b22a90c188", "ctime": 1722475475.8994846, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 488636587, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1722475475.598484, "nlink": 1, "path": "/etc/containers/systemd/envoy-proxy-configmap.yml", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 2102, "uid": 0, "version": "1134094097", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Wednesday 31 July 2024 21:25:42 -0400 (0:00:00.358) 0:01:56.154 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Wednesday 31 July 2024 21:25:42 -0400 (0:00:00.095) 0:01:56.250 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Wednesday 31 July 2024 21:25:43 -0400 (0:00:00.344) 0:01:56.594 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Wednesday 31 July 2024 21:25:43 -0400 (0:00:00.033) 0:01:56.627 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Wednesday 31 July 2024 21:25:43 -0400 (0:00:00.037) 0:01:56.665 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Wednesday 31 July 2024 21:25:43 -0400 (0:00:00.030) 0:01:56.696 ******** changed: [managed_node1] => { "changed": true, "path": "/etc/containers/systemd/envoy-proxy-configmap.yml", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Wednesday 31 July 2024 21:25:43 -0400 (0:00:00.355) 0:01:57.051 ******** ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Wednesday 31 July 2024 21:25:44 -0400 (0:00:00.584) 0:01:57.635 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95 Wednesday 31 July 2024 21:25:44 -0400 (0:00:00.036) 0:01:57.671 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Wednesday 31 July 2024 21:25:44 -0400 (0:00:00.047) 0:01:57.719 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Wednesday 31 July 2024 21:25:44 -0400 (0:00:00.029) 0:01:57.749 ******** changed: [managed_node1] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.029036", "end": "2024-07-31 21:25:44.629306", "rc": 0, "start": "2024-07-31 21:25:44.600270" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127 Wednesday 31 July 2024 21:25:44 -0400 (0:00:00.422) 0:01:58.171 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:25:44 -0400 (0:00:00.058) 0:01:58.230 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:25:44 -0400 (0:00:00.034) 0:01:58.264 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:25:44 -0400 (0:00:00.031) 0:01:58.296 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137 Wednesday 31 July 2024 21:25:44 -0400 (0:00:00.033) 0:01:58.329 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.030231", "end": "2024-07-31 21:25:45.212967", "rc": 0, "start": "2024-07-31 21:25:45.182736" } STDOUT: quay.io/linux-system-roles/mysql 5.6 dd3b2a5dcb48 2 years ago 308 MB TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146 Wednesday 31 July 2024 21:25:45 -0400 (0:00:00.394) 0:01:58.723 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.031052", "end": "2024-07-31 21:25:45.601248", "rc": 0, "start": "2024-07-31 21:25:45.570196" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155 Wednesday 31 July 2024 21:25:45 -0400 (0:00:00.386) 0:01:59.109 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.031912", "end": "2024-07-31 21:25:45.993050", "rc": 0, "start": "2024-07-31 21:25:45.961138" } STDOUT: ea182df3f7ef quay.io/linux-system-roles/mysql:5.6 mysqld About a minute ago Up About a minute (healthy) quadlet-demo-mysql TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164 Wednesday 31 July 2024 21:25:46 -0400 (0:00:00.395) 0:01:59.505 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.056798", "end": "2024-07-31 21:25:46.420773", "rc": 0, "start": "2024-07-31 21:25:46.363975" } STDOUT: podman podman-default-kube-network systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173 Wednesday 31 July 2024 21:25:46 -0400 (0:00:00.424) 0:01:59.929 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 Wednesday 31 July 2024 21:25:46 -0400 (0:00:00.406) 0:02:00.335 ******** ok: [managed_node1] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cni-dhcp.service": { "name": "cni-dhcp.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnsmasq.service": { "name": "dnsmasq.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ea182df3f7eff7c985466c...d656e421db49137275cc20e2f48b59c608623.service": { "name": "ea182df3f7eff7c985466c...d656e421db49137275cc20e2f48b59c608623.service", "source": "systemd", "state": "inactive", "status": "transient" }, "ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service": { "name": "ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-mysql.service": { "name": "quadlet-demo-mysql.service", "source": "systemd", "state": "running", "status": "generated" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 31 July 2024 21:25:48 -0400 (0:00:01.589) 0:02:01.925 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 31 July 2024 21:25:48 -0400 (0:00:00.032) 0:02:01.958 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Install]\nWantedBy=default.target\n\n[Container]\nImage=quay.io/linux-system-roles/mysql:5.6\nContainerName=quadlet-demo-mysql\nVolume=quadlet-demo-mysql.volume:/var/lib/mysql\nVolume=/tmp/quadlet_demo:/var/lib/quadlet_demo:Z\nNetwork=quadlet-demo.network\nSecret=mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD\nHealthCmd=/bin/true\nHealthOnFailure=kill\n", "__podman_quadlet_template_src": "quadlet-demo-mysql.container.j2" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 31 July 2024 21:25:48 -0400 (0:00:00.103) 0:02:02.062 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 31 July 2024 21:25:48 -0400 (0:00:00.041) 0:02:02.104 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 31 July 2024 21:25:48 -0400 (0:00:00.036) 0:02:02.140 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 31 July 2024 21:25:48 -0400 (0:00:00.044) 0:02:02.185 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:25:48 -0400 (0:00:00.062) 0:02:02.247 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:25:48 -0400 (0:00:00.106) 0:02:02.353 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:25:48 -0400 (0:00:00.041) 0:02:02.395 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:25:49 -0400 (0:00:00.047) 0:02:02.443 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:25:49 -0400 (0:00:00.373) 0:02:02.816 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:25:49 -0400 (0:00:00.044) 0:02:02.860 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:25:49 -0400 (0:00:00.355) 0:02:03.216 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:25:49 -0400 (0:00:00.033) 0:02:03.249 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:25:49 -0400 (0:00:00.033) 0:02:03.283 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:25:49 -0400 (0:00:00.033) 0:02:03.316 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:25:49 -0400 (0:00:00.031) 0:02:03.348 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:25:49 -0400 (0:00:00.034) 0:02:03.383 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:25:49 -0400 (0:00:00.031) 0:02:03.415 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:25:50 -0400 (0:00:00.034) 0:02:03.449 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 31 July 2024 21:25:50 -0400 (0:00:00.073) 0:02:03.523 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 31 July 2024 21:25:50 -0400 (0:00:00.059) 0:02:03.582 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 31 July 2024 21:25:50 -0400 (0:00:00.038) 0:02:03.621 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 31 July 2024 21:25:50 -0400 (0:00:00.033) 0:02:03.655 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_images": [ "quay.io/linux-system-roles/mysql:5.6" ], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.container", "__podman_volumes": [ "/tmp/quadlet_demo" ] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 31 July 2024 21:25:50 -0400 (0:00:00.076) 0:02:03.732 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 31 July 2024 21:25:50 -0400 (0:00:00.037) 0:02:03.769 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Wednesday 31 July 2024 21:25:50 -0400 (0:00:00.078) 0:02:03.848 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Wednesday 31 July 2024 21:25:50 -0400 (0:00:00.032) 0:02:03.880 ******** changed: [managed_node1] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql.service", "state": "stopped", "status": { "ActiveEnterTimestamp": "Wed 2024-07-31 21:24:32 EDT", "ActiveEnterTimestampMonotonic": "294632105", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "quadlet-demo-mysql-volume.service -.mount tmp.mount systemd-journald.socket basic.target quadlet-demo-network.service system.slice sysinit.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-31 21:24:32 EDT", "AssertTimestampMonotonic": "294253914", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-31 21:24:32 EDT", "ConditionTimestampMonotonic": "294253913", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroup": "/system.slice/quadlet-demo-mysql.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "yes", "DelegateControllers": "cpu cpuacct cpuset io blkio memory devices pids", "Description": "quadlet-demo-mysql.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "0-1", "EffectiveMemoryNodes": "0", "Environment": "PODMAN_SYSTEMD_UNIT=quadlet-demo-mysql.service", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "56515", "ExecMainStartTimestamp": "Wed 2024-07-31 21:24:32 EDT", "ExecMainStartTimestampMonotonic": "294632083", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman run --name=quadlet-demo-mysql --cidfile=/run/quadlet-demo-mysql.cid --replace --rm --cgroups=split --network=systemd-quadlet-demo --sdnotify=conmon -d -v systemd-quadlet-demo-mysql:/var/lib/mysql -v /tmp/quadlet_demo:/var/lib/quadlet_demo:Z --secret mysql-root-password-container,type=env,target=MYSQL_ROOT_PASSWORD --health-cmd /bin/true --health-on-failure kill quay.io/linux-system-roles/mysql:5.6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopPost": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman rm -v -f -i --cidfile=/run/quadlet-demo-mysql.cid ; ignore_errors=yes ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-mysql.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-31 21:24:32 EDT", "InactiveExitTimestampMonotonic": "294258499", "InvocationID": "64c36cbc54384a4cb736ec2c37ae570e", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "mixed", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "56515", "MemoryAccounting": "yes", "MemoryCurrent": "604114944", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "all", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "-.mount system.slice quadlet-demo-network.service sysinit.target quadlet-demo-mysql-volume.service", "RequiresMountsFor": "/run/containers /tmp/quadlet_demo", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "SourcePath": "/etc/containers/systemd/quadlet-demo-mysql.container", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Wed 2024-07-31 21:24:32 EDT", "StateChangeTimestampMonotonic": "294632105", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "24", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "notify", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestamp": "Wed 2024-07-31 21:24:32 EDT", "WatchdogTimestampMonotonic": "294632100", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Wednesday 31 July 2024 21:25:53 -0400 (0:00:02.756) 0:02:06.636 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722475471.877476, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "ca62b2ad3cc9afb5b5371ebbf797b9bc4fd7edd4", "ctime": 1722475471.3124747, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 408944795, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1722475471.0624743, "nlink": 1, "path": "/etc/containers/systemd/quadlet-demo-mysql.container", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 363, "uid": 0, "version": "1913325720", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Wednesday 31 July 2024 21:25:53 -0400 (0:00:00.368) 0:02:07.005 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Wednesday 31 July 2024 21:25:53 -0400 (0:00:00.060) 0:02:07.066 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Wednesday 31 July 2024 21:25:53 -0400 (0:00:00.351) 0:02:07.417 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Wednesday 31 July 2024 21:25:54 -0400 (0:00:00.048) 0:02:07.465 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Wednesday 31 July 2024 21:25:54 -0400 (0:00:00.032) 0:02:07.497 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Wednesday 31 July 2024 21:25:54 -0400 (0:00:00.035) 0:02:07.533 ******** changed: [managed_node1] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.container", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Wednesday 31 July 2024 21:25:54 -0400 (0:00:00.369) 0:02:07.902 ******** ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Wednesday 31 July 2024 21:25:55 -0400 (0:00:00.617) 0:02:08.520 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95 Wednesday 31 July 2024 21:25:55 -0400 (0:00:00.417) 0:02:08.938 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Wednesday 31 July 2024 21:25:55 -0400 (0:00:00.048) 0:02:08.986 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Wednesday 31 July 2024 21:25:55 -0400 (0:00:00.031) 0:02:09.018 ******** changed: [managed_node1] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.244077", "end": "2024-07-31 21:25:56.125299", "rc": 0, "start": "2024-07-31 21:25:55.881222" } STDOUT: dd3b2a5dcb48ff61113592ed5ddd762581be4387c7bc552375a2159422aa6bf5 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127 Wednesday 31 July 2024 21:25:56 -0400 (0:00:00.616) 0:02:09.634 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:25:56 -0400 (0:00:00.060) 0:02:09.694 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:25:56 -0400 (0:00:00.032) 0:02:09.727 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:25:56 -0400 (0:00:00.034) 0:02:09.761 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137 Wednesday 31 July 2024 21:25:56 -0400 (0:00:00.031) 0:02:09.793 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.037453", "end": "2024-07-31 21:25:56.687237", "rc": 0, "start": "2024-07-31 21:25:56.649784" } TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146 Wednesday 31 July 2024 21:25:56 -0400 (0:00:00.403) 0:02:10.196 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.033172", "end": "2024-07-31 21:25:57.085163", "rc": 0, "start": "2024-07-31 21:25:57.051991" } STDOUT: local systemd-quadlet-demo-mysql local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155 Wednesday 31 July 2024 21:25:57 -0400 (0:00:00.398) 0:02:10.595 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.032645", "end": "2024-07-31 21:25:57.492188", "rc": 0, "start": "2024-07-31 21:25:57.459543" } TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164 Wednesday 31 July 2024 21:25:57 -0400 (0:00:00.408) 0:02:11.003 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.058416", "end": "2024-07-31 21:25:57.915886", "rc": 0, "start": "2024-07-31 21:25:57.857470" } STDOUT: podman podman-default-kube-network systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173 Wednesday 31 July 2024 21:25:57 -0400 (0:00:00.419) 0:02:11.423 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 Wednesday 31 July 2024 21:25:58 -0400 (0:00:00.395) 0:02:11.818 ******** ok: [managed_node1] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cni-dhcp.service": { "name": "cni-dhcp.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnsmasq.service": { "name": "dnsmasq.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quadlet-demo-mysql-volume.service": { "name": "quadlet-demo-mysql-volume.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 31 July 2024 21:26:00 -0400 (0:00:01.626) 0:02:13.444 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 31 July 2024 21:26:00 -0400 (0:00:00.034) 0:02:13.479 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo-mysql.volume", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Volume]", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 31 July 2024 21:26:00 -0400 (0:00:00.045) 0:02:13.524 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 31 July 2024 21:26:00 -0400 (0:00:00.039) 0:02:13.564 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 31 July 2024 21:26:00 -0400 (0:00:00.033) 0:02:13.597 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo-mysql", "__podman_quadlet_type": "volume", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 31 July 2024 21:26:00 -0400 (0:00:00.047) 0:02:13.644 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:26:00 -0400 (0:00:00.061) 0:02:13.706 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:26:00 -0400 (0:00:00.040) 0:02:13.747 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:26:00 -0400 (0:00:00.041) 0:02:13.788 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:26:00 -0400 (0:00:00.044) 0:02:13.832 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:26:00 -0400 (0:00:00.369) 0:02:14.202 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:26:00 -0400 (0:00:00.041) 0:02:14.244 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.360) 0:02:14.604 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.071) 0:02:14.676 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.034) 0:02:14.711 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.032) 0:02:14.744 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.035) 0:02:14.779 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.031) 0:02:14.811 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.034) 0:02:14.845 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.031) 0:02:14.877 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.034) 0:02:14.911 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-mysql-volume.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.057) 0:02:14.969 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.036) 0:02:15.005 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.031) 0:02:15.037 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo-mysql.volume", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.073) 0:02:15.111 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.040) 0:02:15.151 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.076) 0:02:15.227 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Wednesday 31 July 2024 21:26:01 -0400 (0:00:00.035) 0:02:15.262 ******** changed: [managed_node1] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-mysql-volume.service", "state": "stopped", "status": { "ActiveEnterTimestamp": "Wed 2024-07-31 21:24:20 EDT", "ActiveEnterTimestampMonotonic": "282554903", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "sysinit.target -.mount systemd-journald.socket system.slice basic.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-31 21:24:20 EDT", "AssertTimestampMonotonic": "282506135", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-31 21:24:20 EDT", "ConditionTimestampMonotonic": "282506134", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroup": "/system.slice/quadlet-demo-mysql-volume.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-mysql-volume.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "1", "ExecMainExitTimestamp": "Wed 2024-07-31 21:24:20 EDT", "ExecMainExitTimestampMonotonic": "282554665", "ExecMainPID": "55122", "ExecMainStartTimestamp": "Wed 2024-07-31 21:24:20 EDT", "ExecMainStartTimestampMonotonic": "282506962", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman volume create --ignore systemd-quadlet-demo-mysql ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-mysql-volume.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-mysql-volume.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-31 21:24:20 EDT", "InactiveExitTimestampMonotonic": "282507023", "InvocationID": "b908f6ff799245919e1ab0b199414397", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "0", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-mysql-volume.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "sysinit.target -.mount system.slice", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Wed 2024-07-31 21:24:20 EDT", "StateChangeTimestampMonotonic": "282554903", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "exited", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-mysql-volume", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "0", "TasksMax": "22405", "TimeoutStartUSec": "infinity", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Wednesday 31 July 2024 21:26:02 -0400 (0:00:00.663) 0:02:15.926 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722475460.1394506, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "585f8cbdf0ec73000f9227dcffbef71e9552ea4a", "ctime": 1722475459.5624495, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 253755778, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1722475459.3044488, "nlink": 1, "path": "/etc/containers/systemd/quadlet-demo-mysql.volume", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 9, "uid": 0, "version": "1659680150", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Wednesday 31 July 2024 21:26:02 -0400 (0:00:00.372) 0:02:16.299 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Wednesday 31 July 2024 21:26:02 -0400 (0:00:00.061) 0:02:16.361 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Wednesday 31 July 2024 21:26:03 -0400 (0:00:00.357) 0:02:16.718 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Wednesday 31 July 2024 21:26:03 -0400 (0:00:00.046) 0:02:16.765 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Wednesday 31 July 2024 21:26:03 -0400 (0:00:00.032) 0:02:16.797 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Wednesday 31 July 2024 21:26:03 -0400 (0:00:00.033) 0:02:16.831 ******** changed: [managed_node1] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo-mysql.volume", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Wednesday 31 July 2024 21:26:03 -0400 (0:00:00.367) 0:02:17.199 ******** ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Wednesday 31 July 2024 21:26:04 -0400 (0:00:00.589) 0:02:17.788 ******** changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95 Wednesday 31 July 2024 21:26:04 -0400 (0:00:00.426) 0:02:18.215 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Wednesday 31 July 2024 21:26:04 -0400 (0:00:00.048) 0:02:18.263 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Wednesday 31 July 2024 21:26:04 -0400 (0:00:00.031) 0:02:18.295 ******** changed: [managed_node1] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.037564", "end": "2024-07-31 21:26:05.187222", "rc": 0, "start": "2024-07-31 21:26:05.149658" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127 Wednesday 31 July 2024 21:26:05 -0400 (0:00:00.400) 0:02:18.696 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:26:05 -0400 (0:00:00.059) 0:02:18.755 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:26:05 -0400 (0:00:00.033) 0:02:18.788 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:26:05 -0400 (0:00:00.032) 0:02:18.821 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137 Wednesday 31 July 2024 21:26:05 -0400 (0:00:00.035) 0:02:18.856 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.034012", "end": "2024-07-31 21:26:05.750152", "rc": 0, "start": "2024-07-31 21:26:05.716140" } TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146 Wednesday 31 July 2024 21:26:05 -0400 (0:00:00.447) 0:02:19.303 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.034793", "end": "2024-07-31 21:26:06.201068", "rc": 0, "start": "2024-07-31 21:26:06.166275" } STDOUT: local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155 Wednesday 31 July 2024 21:26:06 -0400 (0:00:00.408) 0:02:19.712 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.036798", "end": "2024-07-31 21:26:06.614183", "rc": 0, "start": "2024-07-31 21:26:06.577385" } TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164 Wednesday 31 July 2024 21:26:06 -0400 (0:00:00.411) 0:02:20.124 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.058887", "end": "2024-07-31 21:26:07.048231", "rc": 0, "start": "2024-07-31 21:26:06.989344" } STDOUT: podman podman-default-kube-network systemd-quadlet-demo TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173 Wednesday 31 July 2024 21:26:07 -0400 (0:00:00.433) 0:02:20.557 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 Wednesday 31 July 2024 21:26:07 -0400 (0:00:00.402) 0:02:20.960 ******** ok: [managed_node1] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cni-dhcp.service": { "name": "cni-dhcp.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnsmasq.service": { "name": "dnsmasq.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quadlet-demo-network.service": { "name": "quadlet-demo-network.service", "source": "systemd", "state": "stopped", "status": "generated" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 31 July 2024 21:26:09 -0400 (0:00:01.582) 0:02:22.542 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Wednesday 31 July 2024 21:26:09 -0400 (0:00:00.064) 0:02:22.607 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_file_src": "quadlet-demo.network", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "[Network]\nSubnet=192.168.30.0/24\nGateway=192.168.30.1\nLabel=app=wordpress", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Wednesday 31 July 2024 21:26:09 -0400 (0:00:00.047) 0:02:22.654 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Wednesday 31 July 2024 21:26:09 -0400 (0:00:00.041) 0:02:22.695 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Wednesday 31 July 2024 21:26:09 -0400 (0:00:00.033) 0:02:22.729 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_name": "quadlet-demo", "__podman_quadlet_type": "network", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Wednesday 31 July 2024 21:26:09 -0400 (0:00:00.044) 0:02:22.774 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Wednesday 31 July 2024 21:26:09 -0400 (0:00:00.062) 0:02:22.836 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Wednesday 31 July 2024 21:26:09 -0400 (0:00:00.041) 0:02:22.878 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Wednesday 31 July 2024 21:26:09 -0400 (0:00:00.040) 0:02:22.919 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get group information] **************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:28 Wednesday 31 July 2024 21:26:09 -0400 (0:00:00.045) 0:02:22.964 ******** ok: [managed_node1] => { "ansible_facts": { "getent_group": { "root": [ "x", "0", "" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Set group name] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:35 Wednesday 31 July 2024 21:26:09 -0400 (0:00:00.377) 0:02:23.341 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_group_name": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:39 Wednesday 31 July 2024 21:26:09 -0400 (0:00:00.043) 0:02:23.385 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722474540.8516936, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "bb5b46ffbafcaa8c4021f3c8b3cb8594f48ef34b", "ctime": 1722474507.0424926, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 6884013, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-sharedlib", "mode": "0755", "mtime": 1700557386.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 12640, "uid": 0, "version": "1423826263", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check user with getsubids] ************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:50 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.366) 0:02:23.751 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check group with getsubids] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:55 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.072) 0:02:23.824 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:60 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.034) 0:02:23.859 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:74 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.032) 0:02:23.891 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:79 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.033) 0:02:23.925 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:84 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.031) 0:02:23.956 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:94 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.034) 0:02:23.991 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if group not in subgid file] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:101 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.032) 0:02:24.023 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.032) 0:02:24.055 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "quadlet-demo-network.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.060) 0:02:24.116 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.037) 0:02:24.153 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:87 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.032) 0:02:24.185 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/quadlet-demo.network", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:103 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.071) 0:02:24.257 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:110 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.037) 0:02:24.295 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.075) 0:02:24.371 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Wednesday 31 July 2024 21:26:10 -0400 (0:00:00.032) 0:02:24.403 ******** changed: [managed_node1] => { "changed": true, "enabled": false, "failed_when_result": false, "name": "quadlet-demo-network.service", "state": "stopped", "status": { "ActiveEnterTimestamp": "Wed 2024-07-31 21:24:16 EDT", "ActiveEnterTimestampMonotonic": "278403881", "ActiveExitTimestampMonotonic": "0", "ActiveState": "active", "After": "system.slice systemd-journald.socket sysinit.target -.mount basic.target", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Wed 2024-07-31 21:24:16 EDT", "AssertTimestampMonotonic": "278346423", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Wed 2024-07-31 21:24:16 EDT", "ConditionTimestampMonotonic": "278346422", "ConfigurationDirectoryMode": "0755", "Conflicts": "shutdown.target", "ControlGroup": "/system.slice/quadlet-demo-network.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "quadlet-demo-network.service", "DevicePolicy": "auto", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "1", "ExecMainExitTimestamp": "Wed 2024-07-31 21:24:16 EDT", "ExecMainExitTimestampMonotonic": "278403653", "ExecMainPID": "54213", "ExecMainStartTimestamp": "Wed 2024-07-31 21:24:16 EDT", "ExecMainStartTimestampMonotonic": "278347288", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/bin/podman ; argv[]=/usr/bin/podman network create --ignore --subnet=192.168.30.0/24 --gateway=192.168.30.1 --label app=wordpress systemd-quadlet-demo ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/run/systemd/generator/quadlet-demo-network.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "quadlet-demo-network.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestamp": "Wed 2024-07-31 21:24:16 EDT", "InactiveExitTimestampMonotonic": "278347336", "InvocationID": "7c4f22f32b6640609ec87dc70477eb03", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "MemoryAccounting": "yes", "MemoryCurrent": "0", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "quadlet-demo-network.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "yes", "RemoveIPC": "no", "Requires": "system.slice sysinit.target -.mount", "RequiresMountsFor": "/run/containers", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Wed 2024-07-31 21:24:16 EDT", "StateChangeTimestampMonotonic": "278403881", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "exited", "SuccessAction": "none", "SyslogFacility": "3", "SyslogIdentifier": "quadlet-demo-network", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "0", "TasksMax": "22405", "TimeoutStartUSec": "infinity", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:33 Wednesday 31 July 2024 21:26:11 -0400 (0:00:00.638) 0:02:25.042 ******** ok: [managed_node1] => { "changed": false, "stat": { "atime": 1722475455.9444416, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "e57c08d49aff4bae8daab138d913aeddaa8682a0", "ctime": 1722475455.3784404, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 226492620, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1722475455.0494397, "nlink": 1, "path": "/etc/containers/systemd/quadlet-demo.network", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 74, "uid": 0, "version": "2048984443", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:38 Wednesday 31 July 2024 21:26:11 -0400 (0:00:00.368) 0:02:25.411 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Wednesday 31 July 2024 21:26:12 -0400 (0:00:00.061) 0:02:25.472 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Wednesday 31 July 2024 21:26:12 -0400 (0:00:00.353) 0:02:25.825 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet yaml file] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:44 Wednesday 31 July 2024 21:26:12 -0400 (0:00:00.046) 0:02:25.871 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Reset raw variable] ******************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:52 Wednesday 31 July 2024 21:26:12 -0400 (0:00:00.031) 0:02:25.903 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_raw": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Remove quadlet file] ****************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:42 Wednesday 31 July 2024 21:26:12 -0400 (0:00:00.033) 0:02:25.936 ******** changed: [managed_node1] => { "changed": true, "path": "/etc/containers/systemd/quadlet-demo.network", "state": "absent" } TASK [fedora.linux_system_roles.podman : Refresh systemd] ********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:48 Wednesday 31 July 2024 21:26:12 -0400 (0:00:00.367) 0:02:26.304 ******** ok: [managed_node1] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.podman : Remove managed resource] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:58 Wednesday 31 July 2024 21:26:13 -0400 (0:00:00.586) 0:02:26.891 ******** changed: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } TASK [fedora.linux_system_roles.podman : Remove volumes] *********************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:95 Wednesday 31 July 2024 21:26:13 -0400 (0:00:00.461) 0:02:27.353 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Clear parsed podman variable] ********* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:112 Wednesday 31 July 2024 21:26:13 -0400 (0:00:00.046) 0:02:27.399 ******** ok: [managed_node1] => { "ansible_facts": { "__podman_quadlet_parsed": null }, "changed": false } TASK [fedora.linux_system_roles.podman : Prune images no longer in use] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:116 Wednesday 31 July 2024 21:26:13 -0400 (0:00:00.031) 0:02:27.430 ******** changed: [managed_node1] => { "changed": true, "cmd": [ "podman", "image", "prune", "--all", "-f" ], "delta": "0:00:00.038066", "end": "2024-07-31 21:26:14.331143", "rc": 0, "start": "2024-07-31 21:26:14.293077" } TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:127 Wednesday 31 July 2024 21:26:14 -0400 (0:00:00.410) 0:02:27.840 ******** included: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed_node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Wednesday 31 July 2024 21:26:14 -0400 (0:00:00.057) 0:02:27.898 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Wednesday 31 July 2024 21:26:14 -0400 (0:00:00.035) 0:02:27.933 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Wednesday 31 July 2024 21:26:14 -0400 (0:00:00.030) 0:02:27.964 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : For testing and debugging - images] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:137 Wednesday 31 July 2024 21:26:14 -0400 (0:00:00.033) 0:02:27.997 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "images", "-n" ], "delta": "0:00:00.034284", "end": "2024-07-31 21:26:14.888320", "rc": 0, "start": "2024-07-31 21:26:14.854036" } TASK [fedora.linux_system_roles.podman : For testing and debugging - volumes] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:146 Wednesday 31 July 2024 21:26:14 -0400 (0:00:00.436) 0:02:28.434 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "volume", "ls", "-n" ], "delta": "0:00:00.035997", "end": "2024-07-31 21:26:15.328209", "rc": 0, "start": "2024-07-31 21:26:15.292212" } STDOUT: local wp-pv-claim local envoy-proxy-config local envoy-certificates TASK [fedora.linux_system_roles.podman : For testing and debugging - containers] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:155 Wednesday 31 July 2024 21:26:15 -0400 (0:00:00.401) 0:02:28.836 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "ps", "--noheading" ], "delta": "0:00:00.036075", "end": "2024-07-31 21:26:15.754084", "rc": 0, "start": "2024-07-31 21:26:15.718009" } TASK [fedora.linux_system_roles.podman : For testing and debugging - networks] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:164 Wednesday 31 July 2024 21:26:15 -0400 (0:00:00.425) 0:02:29.261 ******** ok: [managed_node1] => { "changed": false, "cmd": [ "podman", "network", "ls", "-n", "-q" ], "delta": "0:00:00.049359", "end": "2024-07-31 21:26:16.172964", "rc": 0, "start": "2024-07-31 21:26:16.123605" } STDOUT: podman podman-default-kube-network TASK [fedora.linux_system_roles.podman : For testing and debugging - secrets] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:173 Wednesday 31 July 2024 21:26:16 -0400 (0:00:00.422) 0:02:29.684 ******** ok: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : For testing and debugging - services] *** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 Wednesday 31 July 2024 21:26:16 -0400 (0:00:00.399) 0:02:30.084 ******** ok: [managed_node1] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "running", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "certmonger.service": { "name": "certmonger.service", "source": "systemd", "state": "running", "status": "enabled" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cni-dhcp.service": { "name": "cni-dhcp.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.fedoraproject.FirewallD1.service": { "name": "dbus-org.fedoraproject.FirewallD1.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "active", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnsmasq.service": { "name": "dnsmasq.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "running", "status": "enabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "podman-auto-update.service": { "name": "podman-auto-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-clean-transient.service": { "name": "podman-clean-transient.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman-kube@.service": { "name": "podman-kube@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "podman-restart.service": { "name": "podman-restart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "podman.service": { "name": "podman.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:114 Wednesday 31 July 2024 21:26:18 -0400 (0:00:01.636) 0:02:31.720 ******** skipping: [managed_node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:143 Wednesday 31 July 2024 21:26:18 -0400 (0:00:00.031) 0:02:31.751 ******** TASK [fedora.linux_system_roles.podman : Handle credential files - absent] ***** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:149 Wednesday 31 July 2024 21:26:18 -0400 (0:00:00.064) 0:02:31.816 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ******** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:158 Wednesday 31 July 2024 21:26:18 -0400 (0:00:00.032) 0:02:31.849 ******** skipping: [managed_node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [Ensure no resources] ***************************************************** task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:176 Wednesday 31 July 2024 21:26:18 -0400 (0:00:00.033) 0:02:31.882 ******** fatal: [managed_node1]: FAILED! => { "assertion": "__podman_test_debug_volumes.stdout == \"\"", "changed": false, "evaluated_to": false } MSG: Assertion failed TASK [Debug] ******************************************************************* task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:187 Wednesday 31 July 2024 21:26:18 -0400 (0:00:00.039) 0:02:31.922 ******** ok: [managed_node1] => { "changed": false, "cmd": "exec 1>&2\nset -x\nset -o pipefail\nsystemctl list-units --plain -l --all | grep quadlet || :\nsystemctl list-unit-files --all | grep quadlet || :\nsystemctl list-units --plain --failed -l --all | grep quadlet || :\n", "delta": "0:00:00.335037", "end": "2024-07-31 21:26:19.107434", "rc": 0, "start": "2024-07-31 21:26:18.772397" } STDERR: + set -o pipefail + systemctl list-units --plain -l --all + grep quadlet + : + systemctl list-unit-files --all + grep quadlet + : + systemctl list-units --plain --failed -l --all + grep quadlet + : TASK [Get journald] ************************************************************ task path: /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:197 Wednesday 31 July 2024 21:26:19 -0400 (0:00:00.692) 0:02:32.615 ******** fatal: [managed_node1]: FAILED! => { "changed": false, "cmd": [ "journalctl", "-ex" ], "delta": "0:00:00.092622", "end": "2024-07-31 21:26:19.560267", "failed_when_result": true, "rc": 0, "start": "2024-07-31 21:26:19.467645" } STDOUT: -- Logs begin at Wed 2024-07-31 21:19:37 EDT, end at Wed 2024-07-31 21:26:19 EDT. -- Jul 31 21:20:54 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[13932]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:20:54 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[14115]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-myxevqxsoyrwxrnoyjsveqikfdnndnum ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475254.503974-23728-171260245378049/AnsiballZ_command.py' Jul 31 21:20:54 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[14115]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:20:54 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-14120.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:20:54 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[14115]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:20:55 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[14248]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-olickqwuvdnojkhlruvwkvmvhvjvhtfk ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475254.9152694-23737-63819979853780/AnsiballZ_service_facts.py' Jul 31 21:20:55 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[14248]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:20:55 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[14251]: ansible-service_facts Invoked Jul 31 21:20:56 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[14248]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:20:57 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[14462]: ansible-getent Invoked with database=group key=1111 fail_key=False service=None split=None Jul 31 21:20:57 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[14586]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:20:57 ip-10-31-45-139.us-east-1.aws.redhat.com chronyd[588]: Selected source 66.205.249.28 (2.centos.pool.ntp.org) Jul 31 21:20:58 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[14711]: ansible-command Invoked with _raw_params=getsubids user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:20:58 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[14835]: ansible-command Invoked with _raw_params=getsubids -g user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:20:59 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[14959]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:20:59 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15084]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-kqebbgjwfvdionkmuhdcdcnnncwjgwuo ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475259.4869573-23835-237314870475316/AnsiballZ_systemd.py' Jul 31 21:20:59 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15084]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:20:59 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[15087]: ansible-systemd Invoked with name=quadlet-basic-unused-volume-volume.service scope=user state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:20:59 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Reloading. Jul 31 21:20:59 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopped quadlet-basic-unused-volume-volume.service. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:20:59 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15084]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:00 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[15224]: ansible-stat Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-unused-volume.volume follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:01 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[15472]: ansible-file Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-unused-volume.volume state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:21:01 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15595]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ixnwmgwicyycrpsvyituljkxyzmuzzbu ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475261.3272684-23879-95833795480902/AnsiballZ_systemd.py' Jul 31 21:21:01 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15595]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:01 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[15598]: ansible-systemd Invoked with daemon_reload=True scope=user daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:21:01 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Reloading. Jul 31 21:21:01 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15595]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:01 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15732]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ilhzvtvycvcraqiooyhomxmconlnvuvq ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475261.8489242-23888-97047916663308/AnsiballZ_command.py' Jul 31 21:21:01 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15732]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-15737.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:02 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15732]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:02 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15865]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-fijiqedksyzmmfjnhtjlkqvjcqsyvdae ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475262.3660257-23901-22659003857520/AnsiballZ_command.py' Jul 31 21:21:02 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15865]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:02 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[15868]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-15870.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:02 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15865]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:03 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15998]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-uwdtemxpsfktrmmeunxapdubguynumxt ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475262.955352-23918-179310943302855/AnsiballZ_command.py' Jul 31 21:21:03 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15998]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:03 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[16001]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:03 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-16003.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:03 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[15998]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:03 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16131]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-atojrlpwkkepltetmxyohncwaohbwiez ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475263.3712122-23927-54855348619309/AnsiballZ_command.py' Jul 31 21:21:03 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16131]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:03 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[16134]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:03 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-16136.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:03 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16131]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:03 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16264]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-qsetbmbiayoiakqbirlrinsltjyqhlcc ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475263.7945569-23936-224243240103360/AnsiballZ_command.py' Jul 31 21:21:03 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16264]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:04 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[16267]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-16269.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:04 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16264]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:04 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16397]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-puhwcygiegpmvlwbbetxmmfzknlsolns ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475264.2128356-23945-76815982155454/AnsiballZ_command.py' Jul 31 21:21:04 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16397]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:04 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[16400]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-16402.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:04 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16397]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:04 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16581]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-wbcykffrtfclzeznmdsxxxmbuzwkbvyt ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475264.656998-23954-170150113327032/AnsiballZ_command.py' Jul 31 21:21:04 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16581]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-16586.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:04 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16581]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:05 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16714]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-mgnmlbpehwhbiytejkwzsagmuklsdied ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475265.0711956-23963-191140866663849/AnsiballZ_service_facts.py' Jul 31 21:21:05 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16714]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:05 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[16717]: ansible-service_facts Invoked Jul 31 21:21:06 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[16714]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:07 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[16928]: ansible-getent Invoked with database=group key=1111 fail_key=False service=None split=None Jul 31 21:21:07 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[17052]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:08 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[17177]: ansible-command Invoked with _raw_params=getsubids user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:08 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[17301]: ansible-command Invoked with _raw_params=getsubids -g user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:09 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[17425]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:09 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[17550]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-giswqlyyvdfhidctieszcovpgxohlvlf ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475269.445322-24061-140384314487885/AnsiballZ_systemd.py' Jul 31 21:21:09 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[17550]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:09 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[17553]: ansible-systemd Invoked with name=quadlet-basic-mysql-volume.service scope=user state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:21:09 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Reloading. Jul 31 21:21:09 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopped quadlet-basic-mysql-volume.service. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:09 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[17550]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:10 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[17690]: ansible-stat Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-mysql.volume follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:11 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[17938]: ansible-file Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-mysql.volume state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:21:11 ip-10-31-45-139.us-east-1.aws.redhat.com auditd[532]: Audit daemon rotating log files Jul 31 21:21:11 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18061]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-hcponmqtxviykqssijnzxcoksntlfmne ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475271.282623-24105-265857885287136/AnsiballZ_systemd.py' Jul 31 21:21:11 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18061]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:11 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[18064]: ansible-systemd Invoked with daemon_reload=True scope=user daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:21:11 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Reloading. Jul 31 21:21:11 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18061]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:11 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18198]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-fiqqfrdnmkflqwhqpatwwoavkoyshvwm ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475271.8033707-24114-81641251737561/AnsiballZ_command.py' Jul 31 21:21:11 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18198]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:12 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-18203.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:12 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18198]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:12 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18331]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-vtcsrwjdwsejnvjqadqxlpccqcyodlis ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475272.3290458-24127-159146243047508/AnsiballZ_command.py' Jul 31 21:21:12 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18331]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:12 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[18334]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:12 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-18336.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:12 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18331]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:13 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18464]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-gitiigjwkmlhufsppxgwmgmohlrwhdmw ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475272.9575117-24144-196601070397807/AnsiballZ_command.py' Jul 31 21:21:13 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18464]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:13 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[18467]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:13 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-18469.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:13 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18464]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:13 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18597]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-lazdctdbpgxpzginkrdaelycduwwwvyp ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475273.3704453-24153-209027668478354/AnsiballZ_command.py' Jul 31 21:21:13 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18597]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:13 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[18600]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:13 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-18602.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:13 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18597]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:13 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18730]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-xtbdkdqrctmvljljmyvifqrdrahzuirz ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475273.793907-24162-110202051750787/AnsiballZ_command.py' Jul 31 21:21:13 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18730]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:14 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[18733]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:14 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-18735.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:14 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18730]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:14 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18863]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-guvrefsmhvkskaroqvztgghebagjlqpc ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475274.2112997-24171-217007143218051/AnsiballZ_command.py' Jul 31 21:21:14 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18863]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:14 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[18866]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:14 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-18868.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:14 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[18863]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:14 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[19046]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-xwumxodypaevvnpqicjsqxufmmowlgme ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475274.6474783-24180-110233644924085/AnsiballZ_command.py' Jul 31 21:21:14 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[19046]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:14 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-19051.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:14 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[19046]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:15 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[19179]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-cxvijnmuijacznuninrqmlzsgjgzfojg ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475275.0508864-24189-117862565971417/AnsiballZ_service_facts.py' Jul 31 21:21:15 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[19179]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:15 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[19182]: ansible-service_facts Invoked Jul 31 21:21:16 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[19179]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:17 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[19393]: ansible-getent Invoked with database=group key=1111 fail_key=False service=None split=None Jul 31 21:21:17 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[19517]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:17 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[19642]: ansible-command Invoked with _raw_params=getsubids user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:18 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[19766]: ansible-command Invoked with _raw_params=getsubids -g user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:19 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[19890]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:19 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20015]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-xzggcckzzrrvwaghrlllucmciankpbfx ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475279.4039435-24287-130422713041987/AnsiballZ_systemd.py' Jul 31 21:21:19 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20015]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:19 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[20018]: ansible-systemd Invoked with name=quadlet-basic-unused-network-network.service scope=user state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:21:19 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Reloading. Jul 31 21:21:19 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopped quadlet-basic-unused-network-network.service. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:19 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20015]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:20 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[20155]: ansible-stat Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-unused-network.network follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:21 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[20403]: ansible-file Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-unused-network.network state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:21:21 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20526]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-qogbpdnbqbbhbssbftubdrznymcjcnwl ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475281.2366426-24331-51915213610278/AnsiballZ_systemd.py' Jul 31 21:21:21 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20526]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:21 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[20529]: ansible-systemd Invoked with daemon_reload=True scope=user daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:21:21 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Reloading. Jul 31 21:21:21 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20526]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:21 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20663]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-lzvqlyyncfhjatritjnjrvtpsairhcbn ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475281.7633054-24340-66145104673723/AnsiballZ_command.py' Jul 31 21:21:21 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20663]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:22 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-20668.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:22 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20663]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:22 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20847]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ogtecsdaxkonnuisbtxlvanfhcyknjta ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475282.286317-24353-124450577497987/AnsiballZ_command.py' Jul 31 21:21:22 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20847]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:22 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[20850]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:22 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-20852.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:22 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20847]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20980]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-dkabdxbxiobqckultjfmzaaagipbestv ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475282.8762438-24370-279085803457091/AnsiballZ_command.py' Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20980]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[20983]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-20985.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[20980]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21113]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-oqxpqywdfwwdrwfqfuvwexbzososxjib ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475283.2978077-24379-180733817768510/AnsiballZ_command.py' Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21113]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[21116]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-21118.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21113]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21247]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-hpvllamfnhrlmmrcehvlqgmeswdqutzv ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475283.7170253-24388-266468677855484/AnsiballZ_command.py' Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21247]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:23 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[21250]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:24 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-21252.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:24 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21247]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:24 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21380]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-cbypqourcgygvcqnnfavetuwhwccgcms ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475284.1352823-24397-211636805453752/AnsiballZ_command.py' Jul 31 21:21:24 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21380]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:24 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[21383]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:24 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-21385.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:24 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21380]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:24 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21538]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-qsyvrqasdogxltnpqhwyxdqkyqeudhgy ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475284.568117-24406-207634973323917/AnsiballZ_command.py' Jul 31 21:21:24 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21538]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:24 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-21543.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:24 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21538]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:25 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21672]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-oougffjfgabnjwkgifacmdztkyjsaavb ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475285.0316148-24415-256962572416762/AnsiballZ_service_facts.py' Jul 31 21:21:25 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21672]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:25 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[21675]: ansible-service_facts Invoked Jul 31 21:21:26 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[21672]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:27 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[21886]: ansible-getent Invoked with database=group key=1111 fail_key=False service=None split=None Jul 31 21:21:27 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[22010]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:27 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[22135]: ansible-command Invoked with _raw_params=getsubids user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:28 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[22259]: ansible-command Invoked with _raw_params=getsubids -g user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:29 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[22383]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:29 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[22508]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-vggfnwpisbojgfdzsfnbljjxpoozvbbx ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475289.4176571-24513-87492678999592/AnsiballZ_systemd.py' Jul 31 21:21:29 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[22508]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:29 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[22511]: ansible-systemd Invoked with name=quadlet-basic-network.service scope=user state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:21:29 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Reloading. Jul 31 21:21:29 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopped quadlet-basic-network.service. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:29 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[22508]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:30 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[22648]: ansible-stat Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic.network follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:31 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[22896]: ansible-file Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic.network state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:21:31 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23019]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-aownnmvpnbikhrbcyabpfzsassninbqg ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475291.2511528-24557-60753601373542/AnsiballZ_systemd.py' Jul 31 21:21:31 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23019]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:31 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[23022]: ansible-systemd Invoked with daemon_reload=True scope=user daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:21:31 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Reloading. Jul 31 21:21:31 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23019]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:31 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23156]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-qoqpocfycjtbrntpqecrqxbhnholvyze ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475291.7656438-24566-103738719484504/AnsiballZ_command.py' Jul 31 21:21:31 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23156]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-23161.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:32 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23156]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:32 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23315]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-mtncyoeavzqyqnrhokoblnrygrgvjnow ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475292.2767296-24579-157811662567800/AnsiballZ_command.py' Jul 31 21:21:32 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23315]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:32 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[23318]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-23320.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:32 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23315]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23448]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-gqxhjanlcuigpmqajrztiflhagzqmtom ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475292.8621426-24596-198292766161583/AnsiballZ_command.py' Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23448]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[23451]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-23453.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23448]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23581]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-wlwzzgzjwvmtqqeptlnvicatqcmovllc ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475293.2849054-24605-89307053113333/AnsiballZ_command.py' Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23581]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[23584]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-23586.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23581]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23714]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-rfbnarepgxchufrdzmlaidgtemgeubfr ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475293.6964636-24614-232013257052102/AnsiballZ_command.py' Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23714]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[23717]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:33 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-23719.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:34 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23714]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:34 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23847]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-uyeqmhzqrqlzdglcgopkuciqidtxzmmi ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475294.1227493-24623-114687611982818/AnsiballZ_command.py' Jul 31 21:21:34 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23847]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:34 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[23850]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:34 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-23852.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:34 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23847]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:34 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23980]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ppepadkqjoizjivcajpgiyzqabvjckvh ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475294.5361545-24632-100319592605903/AnsiballZ_command.py' Jul 31 21:21:34 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23980]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:34 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-23985.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:34 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[23980]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:35 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[24113]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-bseqqffhweuorxcyuqscljfdspjvgvyi ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475294.947157-24641-103581038532772/AnsiballZ_service_facts.py' Jul 31 21:21:35 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[24113]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:35 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[24116]: ansible-service_facts Invoked Jul 31 21:21:36 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[24113]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:36 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[24327]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:37 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[24452]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ldmimqsqhfxcqoylrarznvawnlqvpjuv ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475297.084389-24667-202022633003231/AnsiballZ_podman_container_info.py' Jul 31 21:21:37 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[24452]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:37 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[24455]: ansible-containers.podman.podman_container_info Invoked with executable=podman name=None Jul 31 21:21:37 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-24457.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:37 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[24452]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:37 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[24585]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ksagvwyrclcmvflrhpoohoonuuhlmqdj ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475297.5668826-24676-267801132444592/AnsiballZ_command.py' Jul 31 21:21:37 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[24585]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:37 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[24588]: ansible-command Invoked with _raw_params=podman network ls -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:37 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-24590.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:37 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[24585]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[24718]: root : TTY=pts/0 ; PWD=/root ; USER=user_quadlet_basic ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-qescyylddxsaloyjwuthnxmnopcufvbq ; XDG_RUNTIME_DIR=/run/user/1111 /usr/libexec/platform-python /var/tmp/ansible-tmp-1722475297.9814217-24685-194722163799508/AnsiballZ_command.py' Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[24718]: pam_unix(sudo:session): session opened for user user_quadlet_basic by root(uid=0) Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[24721]: ansible-command Invoked with _raw_params=podman secret ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started podman-24723.scope. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com sudo[24718]: pam_unix(sudo:session): session closed for user user_quadlet_basic Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[24851]: ansible-command Invoked with removes=/var/lib/systemd/linger/user_quadlet_basic _raw_params=loginctl disable-linger user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None stdin=None Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopping User Manager for UID 1111... -- Subject: Unit user@1111.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user@1111.service has begun shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopping podman-pause-31376b91.scope. -- Subject: Unit UNIT has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has begun shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopping D-Bus User Message Bus... -- Subject: Unit UNIT has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has begun shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopped target Default. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopped D-Bus User Message Bus. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopped target Basic System. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopped target Paths. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopped target Sockets. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopped target Timers. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopped Mark boot as successful after the user session has run 2 minutes. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Closed D-Bus User Message Bus Socket. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Stopped podman-pause-31376b91.scope. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Removed slice user.slice. -- Subject: Unit UNIT has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Reached target Shutdown. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Started Exit the Session. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[2507]: Reached target Exit the Session. -- Subject: Unit UNIT has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit UNIT has finished starting up. -- -- The start-up result is done. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: user@1111.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit user@1111.service has successfully entered the 'dead' state. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped User Manager for UID 1111. -- Subject: Unit user@1111.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user@1111.service has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopping User runtime directory /run/user/1111... -- Subject: Unit user-runtime-dir@1111.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user-runtime-dir@1111.service has begun shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: run-user-1111.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-user-1111.mount has successfully entered the 'dead' state. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: user-runtime-dir@1111.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit user-runtime-dir@1111.service has successfully entered the 'dead' state. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped User runtime directory /run/user/1111. -- Subject: Unit user-runtime-dir@1111.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user-runtime-dir@1111.service has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Removed slice User Slice of UID 1111. -- Subject: Unit user-1111.slice has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit user-1111.slice has finished shutting down. Jul 31 21:21:38 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: user-1111.slice: Consumed 13.328s CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit user-1111.slice completed and consumed the indicated resources. Jul 31 21:21:39 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[24976]: ansible-command Invoked with _raw_params=loginctl show-user -P State user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:39 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[25100]: ansible-systemd Invoked with name=systemd-logind state=stopped daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None scope=None Jul 31 21:21:39 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopping Login Service... -- Subject: Unit systemd-logind.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-logind.service has begun shutting down. Jul 31 21:21:39 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: systemd-logind.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit systemd-logind.service has successfully entered the 'dead' state. Jul 31 21:21:39 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped Login Service. -- Subject: Unit systemd-logind.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit systemd-logind.service has finished shutting down. Jul 31 21:21:39 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[25228]: ansible-command Invoked with _raw_params=loginctl show-user -P State user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:40 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[25352]: ansible-stat Invoked with path=/var/lib/systemd/linger/user_quadlet_basic follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:42 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[25600]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:21:43 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[25729]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Jul 31 21:21:43 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[25853]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:21:44 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[25977]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:46 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-compat619969666-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-compat619969666-merged.mount has successfully entered the 'dead' state. Jul 31 21:21:46 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-metacopy\x2dcheck492534682-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-metacopy\x2dcheck492534682-merged.mount has successfully entered the 'dead' state. Jul 31 21:21:47 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:21:48 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[26403]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:21:48 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[26527]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:49 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[26652]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:21:50 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[26775]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-basic.network follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Jul 31 21:21:50 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[26874]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1722475309.9515831-25017-182552785010752/source dest=/etc/containers/systemd/quadlet-basic.network owner=root group=0 mode=0644 _original_basename=quadlet-basic.network follow=False checksum=19c9b17be2af9b9deca5c3bd327f048966750682 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Jul 31 21:21:51 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[26999]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:21:51 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:21:51 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:21:51 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:21:51 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[27152]: ansible-systemd Invoked with name=quadlet-basic-network.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Jul 31 21:21:51 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-basic-network.service... -- Subject: Unit quadlet-basic-network.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-network.service has begun starting up. Jul 31 21:21:51 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-basic-network[27159]: quadlet-basic-name Jul 31 21:21:51 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:21:51 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started quadlet-basic-network.service. -- Subject: Unit quadlet-basic-network.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-network.service has finished starting up. -- -- The start-up result is done. Jul 31 21:21:52 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[27312]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:21:53 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[27436]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:54 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[27561]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:21:54 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[27684]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-basic-unused-network.network follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Jul 31 21:21:55 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[27783]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1722475314.4730124-25148-149259057059342/source dest=/etc/containers/systemd/quadlet-basic-unused-network.network owner=root group=0 mode=0644 follow=False _original_basename=systemd.j2 checksum=52c9d75ecaf81203cc1f1a3b1dd00fcd25067b01 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Jul 31 21:21:55 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[27908]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:21:55 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:21:55 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:21:55 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:21:56 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[28061]: ansible-systemd Invoked with name=quadlet-basic-unused-network-network.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Jul 31 21:21:56 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-basic-unused-network-network.service... -- Subject: Unit quadlet-basic-unused-network-network.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-unused-network-network.service has begun starting up. Jul 31 21:21:56 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-basic-unused-network-network[28068]: systemd-quadlet-basic-unused-network Jul 31 21:21:56 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:21:56 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started quadlet-basic-unused-network-network.service. -- Subject: Unit quadlet-basic-unused-network-network.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-unused-network-network.service has finished starting up. -- -- The start-up result is done. Jul 31 21:21:57 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[28247]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:21:57 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[28371]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:21:58 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[28496]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:21:59 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[28619]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-basic-mysql.volume follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Jul 31 21:21:59 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[28718]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1722475318.9277935-25275-193054297512959/source dest=/etc/containers/systemd/quadlet-basic-mysql.volume owner=root group=0 mode=0644 follow=False _original_basename=systemd.j2 checksum=90a3571bfc7670328fe3f8fb625585613dbd9c4a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Jul 31 21:21:59 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[28843]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:21:59 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:22:00 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:22:00 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:22:00 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[28996]: ansible-systemd Invoked with name=quadlet-basic-mysql-volume.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Jul 31 21:22:00 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-basic-mysql-volume.service... -- Subject: Unit quadlet-basic-mysql-volume.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-mysql-volume.service has begun starting up. Jul 31 21:22:00 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-basic-mysql-volume[29003]: quadlet-basic-mysql-name Jul 31 21:22:00 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:22:00 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started quadlet-basic-mysql-volume.service. -- Subject: Unit quadlet-basic-mysql-volume.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-mysql-volume.service has finished starting up. -- -- The start-up result is done. Jul 31 21:22:01 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[29131]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:22:01 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[29255]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:03 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[29380]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:22:03 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[29503]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-basic-unused-volume.volume follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Jul 31 21:22:03 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[29602]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1722475323.4139912-25402-109453500847575/source dest=/etc/containers/systemd/quadlet-basic-unused-volume.volume owner=root group=0 mode=0644 follow=False _original_basename=systemd.j2 checksum=fd0ae560360afa5541b866560b1e849d25e216ef backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Jul 31 21:22:04 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[29727]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:22:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:22:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:22:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:22:05 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[29880]: ansible-systemd Invoked with name=quadlet-basic-unused-volume-volume.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Jul 31 21:22:05 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-basic-unused-volume-volume.service... -- Subject: Unit quadlet-basic-unused-volume-volume.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-unused-volume-volume.service has begun starting up. Jul 31 21:22:05 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-basic-unused-volume-volume[29887]: systemd-quadlet-basic-unused-volume Jul 31 21:22:05 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:22:05 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started quadlet-basic-unused-volume-volume.service. -- Subject: Unit quadlet-basic-unused-volume-volume.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-unused-volume-volume.service has finished starting up. -- -- The start-up result is done. Jul 31 21:22:05 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[30015]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:22:06 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[30139]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:07 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:22:14 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:22:14 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:22:14 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[30484]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:22:15 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[30607]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-basic-mysql.container follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Jul 31 21:22:15 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[30706]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1722475334.8326395-25536-17541919133742/source dest=/etc/containers/systemd/quadlet-basic-mysql.container owner=root group=0 mode=0644 follow=False _original_basename=systemd.j2 checksum=0b6cac7929623f1059e78ef39b8b0a25169b28a6 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Jul 31 21:22:15 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[30831]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:22:15 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:22:15 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:22:15 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[30984]: ansible-systemd Invoked with name=quadlet-basic-mysql.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-basic-mysql.service... -- Subject: Unit quadlet-basic-mysql.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-mysql.service has begun starting up. Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd-udevd[31081]: Using default interface naming scheme 'rhel-8.0'. Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): vethd8acb714: link is not ready Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd8acb714) entered blocking state Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd8acb714) entered disabled state Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: device vethd8acb714 entered promiscuous mode Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd8acb714) entered blocking state Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd8acb714) entered forwarding state Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6199] manager: (cni-podman2): new Bridge device (/org/freedesktop/NetworkManager/Devices/3) Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6210] manager: (vethd8acb714): new Veth device (/org/freedesktop/NetworkManager/Devices/4) Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd8acb714) entered disabled state Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): eth0: link is not ready Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd-udevd[31081]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable. Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd-udevd[31081]: Could not generate persistent MAC address for cni-podman2: No such file or directory Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd-udevd[31082]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable. Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd-udevd[31082]: Could not generate persistent MAC address for vethd8acb714: No such file or directory Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): vethd8acb714: link becomes ready Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd8acb714) entered blocking state Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd8acb714) entered forwarding state Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6437] device (cni-podman2): carrier: link connected Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6439] device (vethd8acb714): carrier: link connected Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6459] device (cni-podman2): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6464] device (cni-podman2): state change: unavailable -> disconnected (reason 'connection-assumed', sys-iface-state: 'external') Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6470] device (cni-podman2): Activation: starting connection 'cni-podman2' (81426242-2164-454f-9279-997f42fbe551) Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6471] device (cni-podman2): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'external') Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6473] device (cni-podman2): state change: prepare -> config (reason 'none', sys-iface-state: 'external') Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6475] device (cni-podman2): state change: config -> ip-config (reason 'none', sys-iface-state: 'external') Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6476] device (cni-podman2): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'external') Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com dbus-daemon[581]: [system] Activating via systemd: service name='org.freedesktop.nm_dispatcher' unit='dbus-org.freedesktop.nm-dispatcher.service' requested by ':1.3' (uid=0 pid=670 comm="/usr/sbin/NetworkManager --no-daemon " label="system_u:system_r:NetworkManager_t:s0") Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Script Dispatcher Service... -- Subject: Unit NetworkManager-dispatcher.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has begun starting up. Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com dbus-daemon[581]: [system] Successfully activated service 'org.freedesktop.nm_dispatcher' Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started Network Manager Script Dispatcher Service. -- Subject: Unit NetworkManager-dispatcher.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has finished starting up. -- -- The start-up result is done. Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6910] device (cni-podman2): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'external') Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6914] device (cni-podman2): state change: secondaries -> activated (reason 'none', sys-iface-state: 'external') Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475336.6918] device (cni-podman2): Activation: successful, device activated. Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[31157]: listening on cni-podman2(#3): 192.168.29.1 Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[31169]: started, version 2.79 cachesize 150 Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[31169]: compile time options: IPv6 GNU-getopt DBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth DNSSEC loop-detect inotify Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[31169]: using local addresses only for domain dns.podman Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[31169]: reading /etc/resolv.conf Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[31169]: using local addresses only for domain dns.podman Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[31169]: using nameserver 10.29.169.13#53 Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[31169]: using nameserver 10.29.170.12#53 Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[31169]: using nameserver 10.2.32.1#53 Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[31169]: read /run/containers/cni/dnsname/quadlet-basic-name/addnhosts - 1 addresses Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: Adding Red Hat flag eBPF/cgroup. Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started quadlet-basic-mysql.service. -- Subject: Unit quadlet-basic-mysql.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-mysql.service has finished starting up. -- -- The start-up result is done. Jul 31 21:22:16 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-basic-mysql[30991]: 5aa8ee9a1117541d65b612385484e38f160b73d54eca65d8b9f4edd4209b3b49 Jul 31 21:22:17 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[31373]: ansible-command Invoked with _raw_params=cat /etc/containers/systemd/quadlet-basic-mysql.container warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:17 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[31497]: ansible-command Invoked with _raw_params=cat /etc/containers/systemd/quadlet-basic.network warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:18 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[31632]: ansible-command Invoked with _raw_params=cat /etc/containers/systemd/quadlet-basic-mysql.volume warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:18 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[31169]: listening on cni-podman2(#3): fe80::9881:94ff:fe59:33c5%cni-podman2 Jul 31 21:22:18 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[31772]: ansible-command Invoked with _raw_params=podman exec quadlet-basic-mysql-name cat /tmp/test.json warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:21 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[32077]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:21 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[32206]: ansible-getent Invoked with database=passwd key=user_quadlet_basic fail_key=False service=None split=None Jul 31 21:22:22 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[32337]: ansible-getent Invoked with database=group key=1111 fail_key=False service=None split=None Jul 31 21:22:22 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[32482]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:23 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[32616]: ansible-command Invoked with _raw_params=getsubids user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:23 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[32740]: ansible-command Invoked with _raw_params=getsubids -g user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:25 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[32871]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:26 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[32994]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:26 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: NetworkManager-dispatcher.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Jul 31 21:22:27 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[33140]: ansible-getent Invoked with database=group key=1111 fail_key=False service=None split=None Jul 31 21:22:27 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[33264]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:27 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[33389]: ansible-command Invoked with _raw_params=getsubids user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:28 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[33513]: ansible-command Invoked with _raw_params=getsubids -g user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:29 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[33637]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:29 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[33760]: ansible-stat Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-mysql.container follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:30 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[33883]: ansible-file Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-mysql.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:22:31 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[34006]: ansible-getent Invoked with database=group key=1111 fail_key=False service=None split=None Jul 31 21:22:31 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[34130]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:32 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[34255]: ansible-command Invoked with _raw_params=getsubids user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:32 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[34379]: ansible-command Invoked with _raw_params=getsubids -g user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:33 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[34503]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:34 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[34626]: ansible-stat Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-unused-volume.volume follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:34 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[34749]: ansible-file Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-unused-volume.volume state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:22:35 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[34872]: ansible-getent Invoked with database=group key=1111 fail_key=False service=None split=None Jul 31 21:22:36 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[34996]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:36 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[35121]: ansible-command Invoked with _raw_params=getsubids user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:37 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[35245]: ansible-command Invoked with _raw_params=getsubids -g user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:38 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[35369]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:38 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[35492]: ansible-stat Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-mysql.volume follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:38 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[35615]: ansible-file Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-mysql.volume state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:22:40 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[35738]: ansible-getent Invoked with database=group key=1111 fail_key=False service=None split=None Jul 31 21:22:40 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[35862]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:41 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[35987]: ansible-command Invoked with _raw_params=getsubids user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:41 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[36111]: ansible-command Invoked with _raw_params=getsubids -g user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:42 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[36235]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:42 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[36358]: ansible-stat Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-unused-network.network follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:43 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[36481]: ansible-file Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic-unused-network.network state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:22:44 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[36604]: ansible-getent Invoked with database=group key=1111 fail_key=False service=None split=None Jul 31 21:22:45 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[36728]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:45 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[36853]: ansible-command Invoked with _raw_params=getsubids user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:45 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[36977]: ansible-command Invoked with _raw_params=getsubids -g user_quadlet_basic warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:46 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[37101]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:47 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[37224]: ansible-stat Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic.network follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:47 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[37347]: ansible-file Invoked with path=/home/user_quadlet_basic/.config/containers/systemd/quadlet-basic.network state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:22:48 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[37470]: ansible-stat Invoked with path=/run/user/1111 follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:49 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[37593]: ansible-user Invoked with name=user_quadlet_basic uid=1111 state=absent non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on ip-10-31-45-139.us-east-1.aws.redhat.com update_password=always group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None Jul 31 21:22:49 ip-10-31-45-139.us-east-1.aws.redhat.com userdel[37597]: delete user 'user_quadlet_basic' Jul 31 21:22:49 ip-10-31-45-139.us-east-1.aws.redhat.com userdel[37597]: removed group 'user_quadlet_basic' owned by 'user_quadlet_basic' Jul 31 21:22:49 ip-10-31-45-139.us-east-1.aws.redhat.com userdel[37597]: removed shadow group 'user_quadlet_basic' owned by 'user_quadlet_basic' Jul 31 21:22:52 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[37850]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:22:52 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[37979]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Jul 31 21:22:53 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[38103]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:22:53 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[38227]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:57 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[38612]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:22:57 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[38736]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:22:58 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[38861]: ansible-systemd Invoked with name=quadlet-basic-mysql.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:22:58 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:22:58 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:22:58 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:22:59 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopping quadlet-basic-mysql.service... -- Subject: Unit quadlet-basic-mysql.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-mysql.service has begun shutting down. Jul 31 21:23:00 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd8acb714) entered disabled state Jul 31 21:23:00 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: device vethd8acb714 left promiscuous mode Jul 31 21:23:00 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethd8acb714) entered disabled state Jul 31 21:23:00 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: run-netns-netns\x2d21e6b8e7\x2d45fe\x2d26d7\x2d275b\x2dee319bf05989.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-netns-netns\x2d21e6b8e7\x2d45fe\x2d26d7\x2d275b\x2dee319bf05989.mount has successfully entered the 'dead' state. Jul 31 21:23:00 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-5aa8ee9a1117541d65b612385484e38f160b73d54eca65d8b9f4edd4209b3b49-userdata-shm.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay\x2dcontainers-5aa8ee9a1117541d65b612385484e38f160b73d54eca65d8b9f4edd4209b3b49-userdata-shm.mount has successfully entered the 'dead' state. Jul 31 21:23:00 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-0b3d747d4e92392d6746813e8497730ee265b7574335a0da0ca41a1dca639dab-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-0b3d747d4e92392d6746813e8497730ee265b7574335a0da0ca41a1dca639dab-merged.mount has successfully entered the 'dead' state. Jul 31 21:23:00 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-basic-mysql[38897]: 5aa8ee9a1117541d65b612385484e38f160b73d54eca65d8b9f4edd4209b3b49 Jul 31 21:23:00 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:00 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: quadlet-basic-mysql.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-basic-mysql.service has successfully entered the 'dead' state. Jul 31 21:23:00 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-basic-mysql.service. -- Subject: Unit quadlet-basic-mysql.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-mysql.service has finished shutting down. Jul 31 21:23:01 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[39167]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-basic-mysql.container follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:23:02 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[39415]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-basic-mysql.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:23:02 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[39538]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:23:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:23:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:03 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[39822]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:04 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[39953]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:05 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[40083]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:05 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[40213]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:05 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[40343]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:05 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:06 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[40678]: ansible-service_facts Invoked Jul 31 21:23:08 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[40889]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:23:09 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[41013]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:23:10 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[41138]: ansible-systemd Invoked with name=quadlet-basic-unused-volume-volume.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:23:10 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:23:10 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:10 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:10 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: quadlet-basic-unused-volume-volume.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-basic-unused-volume-volume.service has successfully entered the 'dead' state. Jul 31 21:23:10 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-basic-unused-volume-volume.service. -- Subject: Unit quadlet-basic-unused-volume-volume.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-unused-volume-volume.service has finished shutting down. Jul 31 21:23:10 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[41295]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-basic-unused-volume.volume follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:23:11 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[41543]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-basic-unused-volume.volume state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:23:12 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[41666]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:23:12 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:23:12 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:12 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:12 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:13 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[41949]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:13 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[42079]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:14 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[42209]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:14 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:14 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[42339]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:15 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[42470]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:15 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:15 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[42805]: ansible-service_facts Invoked Jul 31 21:23:17 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[43016]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:23:18 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[43140]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:23:19 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[43265]: ansible-systemd Invoked with name=quadlet-basic-mysql-volume.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:23:19 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:23:19 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:19 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:19 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: quadlet-basic-mysql-volume.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-basic-mysql-volume.service has successfully entered the 'dead' state. Jul 31 21:23:19 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-basic-mysql-volume.service. -- Subject: Unit quadlet-basic-mysql-volume.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-mysql-volume.service has finished shutting down. Jul 31 21:23:19 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[43422]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-basic-mysql.volume follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:23:20 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[43670]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-basic-mysql.volume state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:23:21 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[43793]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:23:21 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:23:21 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:21 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:21 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:22 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[44076]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:22 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[44206]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:23 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[44336]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:23 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[44466]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:24 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[44596]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:24 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:24 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:24 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[44932]: ansible-service_facts Invoked Jul 31 21:23:26 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[45143]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:23:27 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[45267]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:23:28 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[45392]: ansible-systemd Invoked with name=quadlet-basic-unused-network-network.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:23:28 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:23:28 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:28 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:28 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: quadlet-basic-unused-network-network.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-basic-unused-network-network.service has successfully entered the 'dead' state. Jul 31 21:23:28 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-basic-unused-network-network.service. -- Subject: Unit quadlet-basic-unused-network-network.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-unused-network-network.service has finished shutting down. Jul 31 21:23:28 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[45549]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-basic-unused-network.network follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:23:29 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[45797]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-basic-unused-network.network state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:23:30 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[45920]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:23:30 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:23:30 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:30 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:30 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:31 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[46279]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:31 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[46409]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:31 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:32 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[46539]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:32 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[46669]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:33 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[46799]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:33 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:34 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[47109]: ansible-service_facts Invoked Jul 31 21:23:35 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[47320]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:23:36 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[47444]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:23:37 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[47569]: ansible-systemd Invoked with name=quadlet-basic-network.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:23:37 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:23:37 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:37 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:37 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: quadlet-basic-network.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-basic-network.service has successfully entered the 'dead' state. Jul 31 21:23:37 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-basic-network.service. -- Subject: Unit quadlet-basic-network.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-basic-network.service has finished shutting down. Jul 31 21:23:38 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[47726]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-basic.network follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:23:38 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[47974]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-basic.network state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:23:39 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[48097]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:23:39 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:23:39 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:39 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:23:40 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:40 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475420.0131] device (cni-podman2): state change: activated -> unmanaged (reason 'unmanaged', sys-iface-state: 'removed') Jul 31 21:23:40 ip-10-31-45-139.us-east-1.aws.redhat.com dbus-daemon[581]: [system] Activating via systemd: service name='org.freedesktop.nm_dispatcher' unit='dbus-org.freedesktop.nm-dispatcher.service' requested by ':1.3' (uid=0 pid=670 comm="/usr/sbin/NetworkManager --no-daemon " label="system_u:system_r:NetworkManager_t:s0") Jul 31 21:23:40 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Script Dispatcher Service... -- Subject: Unit NetworkManager-dispatcher.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has begun starting up. Jul 31 21:23:40 ip-10-31-45-139.us-east-1.aws.redhat.com dbus-daemon[581]: [system] Successfully activated service 'org.freedesktop.nm_dispatcher' Jul 31 21:23:40 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started Network Manager Script Dispatcher Service. -- Subject: Unit NetworkManager-dispatcher.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has finished starting up. -- -- The start-up result is done. Jul 31 21:23:40 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[48441]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:41 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[48571]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:41 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:41 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[48701]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:41 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[48831]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:42 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[48961]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:42 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:42 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:23:43 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[49247]: ansible-service_facts Invoked Jul 31 21:23:47 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[49496]: ansible-setup Invoked with gather_subset=['all'] gather_timeout=10 filter=* fact_path=/etc/ansible/facts.d Jul 31 21:23:47 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[49644]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:23:48 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[49767]: ansible-dnf Invoked with name=['python3-pyasn1', 'python3-cryptography', 'python3-dbus'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Jul 31 21:23:50 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: NetworkManager-dispatcher.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Jul 31 21:23:51 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[49890]: ansible-dnf Invoked with name=['certmonger'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Jul 31 21:23:53 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[50013]: ansible-file Invoked with name=/etc/certmonger//pre-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//pre-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:23:54 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[50136]: ansible-file Invoked with name=/etc/certmonger//post-scripts owner=root group=root mode=0700 state=directory path=/etc/certmonger//post-scripts recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:23:54 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[50259]: ansible-systemd Invoked with name=certmonger state=started enabled=True daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None scope=None Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[50386]: ansible-fedora.linux_system_roles.certificate_request Invoked with name=quadlet_demo dns=['localhost'] directory=/etc/pki/tls wait=True ca=self-sign __header=# # Ansible managed # # system_role:certificate provider_config_directory=/etc/certmonger provider=certmonger key_usage=['digitalSignature', 'keyEncipherment'] extended_key_usage=['id-kp-serverAuth', 'id-kp-clientAuth'] auto_renew=True ip=None email=None common_name=None country=None state=None locality=None organization=None organizational_unit=None contact_email=None key_size=None owner=None group=None mode=None principal=None run_before=None run_after=None Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[50402]: Certificate in file "/etc/pki/tls/certs/quadlet_demo.crt" issued by CA and saved. Jul 31 21:23:55 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:55 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:56 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[50524]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Jul 31 21:23:56 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[50647]: ansible-slurp Invoked with path=/etc/pki/tls/private/quadlet_demo.key src=/etc/pki/tls/private/quadlet_demo.key Jul 31 21:23:56 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[50770]: ansible-slurp Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt src=/etc/pki/tls/certs/quadlet_demo.crt Jul 31 21:23:57 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[50893]: ansible-command Invoked with _raw_params=getcert stop-tracking -f /etc/pki/tls/certs/quadlet_demo.crt warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:23:57 ip-10-31-45-139.us-east-1.aws.redhat.com certmonger[675]: 2024-07-31 21:23:57 [675] Wrote to /var/lib/certmonger/requests/20240801012355 Jul 31 21:23:57 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[51017]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:23:58 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[51140]: ansible-file Invoked with path=/etc/pki/tls/private/quadlet_demo.key state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:23:58 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[51263]: ansible-file Invoked with path=/etc/pki/tls/certs/quadlet_demo.crt state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:23:58 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[51386]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:24:01 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[51634]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:24:01 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[51762]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Jul 31 21:24:02 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[51886]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:24:02 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[52010]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:24:04 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[52135]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:24:04 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[52258]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:24:05 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[52381]: ansible-dnf Invoked with name=['firewalld'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Jul 31 21:24:07 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[52504]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False no_block=False state=None enabled=None force=None user=None scope=None Jul 31 21:24:08 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[52631]: ansible-systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None scope=None Jul 31 21:24:08 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[52758]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Jul 31 21:24:09 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[52881]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Jul 31 21:24:10 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:24:10 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:24:11 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:24:12 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:24:12 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:24:13 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[53457]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:24:13 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[53581]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:24:14 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[53706]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:24:15 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[53829]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.network follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Jul 31 21:24:15 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[53928]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1722475454.7581224-28360-69997062430020/source dest=/etc/containers/systemd/quadlet-demo.network owner=root group=0 mode=0644 _original_basename=quadlet-demo.network follow=False checksum=e57c08d49aff4bae8daab138d913aeddaa8682a0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Jul 31 21:24:15 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[54053]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:24:15 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:24:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:24:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:24:16 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[54206]: ansible-systemd Invoked with name=quadlet-demo-network.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Jul 31 21:24:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-demo-network.service... -- Subject: Unit quadlet-demo-network.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-network.service has begun starting up. Jul 31 21:24:16 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo-network[54213]: systemd-quadlet-demo Jul 31 21:24:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:24:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started quadlet-demo-network.service. -- Subject: Unit quadlet-demo-network.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-network.service has finished starting up. -- -- The start-up result is done. Jul 31 21:24:17 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[54366]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:24:17 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[54490]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:24:18 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[54615]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:24:19 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[54738]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Jul 31 21:24:19 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[54837]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1722475459.0135796-28487-68246258164666/source dest=/etc/containers/systemd/quadlet-demo-mysql.volume owner=root group=0 mode=0644 _original_basename=quadlet-demo-mysql.volume follow=False checksum=585f8cbdf0ec73000f9227dcffbef71e9552ea4a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Jul 31 21:24:20 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[54962]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:24:20 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:24:20 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:24:20 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:24:20 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[55115]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Jul 31 21:24:20 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-demo-mysql-volume.service... -- Subject: Unit quadlet-demo-mysql-volume.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql-volume.service has begun starting up. Jul 31 21:24:20 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo-mysql-volume[55122]: systemd-quadlet-demo-mysql Jul 31 21:24:20 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:24:20 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started quadlet-demo-mysql-volume.service. -- Subject: Unit quadlet-demo-mysql-volume.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql-volume.service has finished starting up. -- -- The start-up result is done. Jul 31 21:24:21 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[55251]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:24:21 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[55375]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:24:23 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[55500]: ansible-file Invoked with path=/tmp/quadlet_demo state=directory owner=root group=root mode=0777 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:24:30 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:24:30 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:24:30 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[55843]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:24:31 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[55966]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Jul 31 21:24:31 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[56065]: ansible-copy Invoked with dest=/etc/containers/systemd/quadlet-demo-mysql.container owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1722475470.7698424-28631-129284827082904/source _original_basename=tmpzdkraay1 follow=False checksum=ca62b2ad3cc9afb5b5371ebbf797b9bc4fd7edd4 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Jul 31 21:24:31 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[56190]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:24:31 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:24:31 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:24:31 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[56343]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-demo-mysql.service... -- Subject: Unit quadlet-demo-mysql.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql.service has begun starting up. Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd-udevd[56414]: Using default interface naming scheme 'rhel-8.0'. Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): vethb66cc4ad: link is not ready Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethb66cc4ad) entered blocking state Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethb66cc4ad) entered disabled state Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: device vethb66cc4ad entered promiscuous mode Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethb66cc4ad) entered blocking state Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethb66cc4ad) entered forwarding state Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.5515] manager: (cni-podman2): new Bridge device (/org/freedesktop/NetworkManager/Devices/5) Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.5529] manager: (vethb66cc4ad): new Veth device (/org/freedesktop/NetworkManager/Devices/6) Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd-udevd[56414]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable. Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd-udevd[56414]: Could not generate persistent MAC address for cni-podman2: No such file or directory Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd-udevd[56415]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable. Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): vethb66cc4ad: link becomes ready Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd-udevd[56415]: Could not generate persistent MAC address for vethb66cc4ad: No such file or directory Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.5731] device (vethb66cc4ad): carrier: link connected Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.5734] device (cni-podman2): carrier: link connected Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.5871] device (cni-podman2): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.5876] device (cni-podman2): state change: unavailable -> disconnected (reason 'connection-assumed', sys-iface-state: 'external') Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.5883] device (cni-podman2): Activation: starting connection 'cni-podman2' (81debe96-1bda-45b3-85f3-ba565e56dd9c) Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.5884] device (cni-podman2): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'external') Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.5886] device (cni-podman2): state change: prepare -> config (reason 'none', sys-iface-state: 'external') Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.5888] device (cni-podman2): state change: config -> ip-config (reason 'none', sys-iface-state: 'external') Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.5889] device (cni-podman2): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'external') Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com dbus-daemon[581]: [system] Activating via systemd: service name='org.freedesktop.nm_dispatcher' unit='dbus-org.freedesktop.nm-dispatcher.service' requested by ':1.3' (uid=0 pid=670 comm="/usr/sbin/NetworkManager --no-daemon " label="system_u:system_r:NetworkManager_t:s0") Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Script Dispatcher Service... -- Subject: Unit NetworkManager-dispatcher.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has begun starting up. Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com dbus-daemon[581]: [system] Successfully activated service 'org.freedesktop.nm_dispatcher' Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started Network Manager Script Dispatcher Service. -- Subject: Unit NetworkManager-dispatcher.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has finished starting up. -- -- The start-up result is done. Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.6249] device (cni-podman2): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'external') Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.6251] device (cni-podman2): state change: secondaries -> activated (reason 'none', sys-iface-state: 'external') Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475472.6256] device (cni-podman2): Activation: successful, device activated. Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56491]: listening on cni-podman2(#5): 192.168.30.1 Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56502]: started, version 2.79 cachesize 150 Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56502]: compile time options: IPv6 GNU-getopt DBus no-i18n IDN2 DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth DNSSEC loop-detect inotify Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56502]: using local addresses only for domain dns.podman Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56502]: reading /etc/resolv.conf Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56502]: using local addresses only for domain dns.podman Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56502]: using nameserver 10.29.169.13#53 Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56502]: using nameserver 10.29.170.12#53 Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56502]: using nameserver 10.2.32.1#53 Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56502]: read /run/containers/cni/dnsname/systemd-quadlet-demo/addnhosts - 1 addresses Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/podman healthcheck run ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623. -- Subject: Unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.timer has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.timer has finished starting up. -- -- The start-up result is done. Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started quadlet-demo-mysql.service. -- Subject: Unit quadlet-demo-mysql.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql.service has finished starting up. -- -- The start-up result is done. Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/podman healthcheck run ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623. -- Subject: Unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service has finished starting up. -- -- The start-up result is done. Jul 31 21:24:32 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo-mysql[56350]: ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623 Jul 31 21:24:33 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service has successfully entered the 'dead' state. Jul 31 21:24:33 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56502]: listening on cni-podman2(#5): fe80::7411:6eff:fe07:8178%cni-podman2 Jul 31 21:24:33 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[56732]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:24:34 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[56867]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:24:35 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[57008]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:24:35 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[57131]: ansible-stat Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Jul 31 21:24:35 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[57230]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1722475475.3095858-28756-53264211105254/source dest=/etc/containers/systemd/envoy-proxy-configmap.yml owner=root group=0 mode=0644 _original_basename=envoy-proxy-configmap.yml follow=False checksum=d681c7d56f912150d041873e880818b22a90c188 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Jul 31 21:24:36 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[57379]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:24:36 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:24:36 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:24:36 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:24:37 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[57532]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:24:37 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[57656]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:24:39 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[57818]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:24:39 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[57941]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.yml follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Jul 31 21:24:40 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[58040]: ansible-copy Invoked with dest=/etc/containers/systemd/quadlet-demo.yml owner=root group=0 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1722475479.3956273-28878-108312669605419/source _original_basename=tmprdlha5az follow=False checksum=998dccde0483b1654327a46ddd89cbaa47650370 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Jul 31 21:24:40 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[58165]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:24:40 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:24:40 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:24:40 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:24:41 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[58325]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:24:42 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[58449]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:24:42 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: NetworkManager-dispatcher.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Jul 31 21:24:42 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[58574]: ansible-slurp Invoked with path=/etc/containers/systemd/quadlet-demo.yml src=/etc/containers/systemd/quadlet-demo.yml Jul 31 21:24:43 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[58720]: ansible-file Invoked with path=/tmp/httpd3 state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:24:43 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[58843]: ansible-file Invoked with path=/tmp/httpd3-create state=directory owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:25:01 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[59430]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:25:02 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[59553]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.kube follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Jul 31 21:25:02 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[59652]: ansible-copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1722475501.7691553-29033-234388092068636/source dest=/etc/containers/systemd/quadlet-demo.kube owner=root group=0 mode=0644 _original_basename=quadlet-demo.kube follow=False checksum=7a5c73a5d935a42431c87bcdbeb8a04ed0909dc7 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None regexp=None delimiter=None Jul 31 21:25:02 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[59777]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:25:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:25:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:03 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[59930]: ansible-systemd Invoked with name=quadlet-demo.service scope=system state=started daemon_reload=False daemon_reexec=False no_block=False enabled=None force=None masked=None user=None Jul 31 21:25:03 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/podman healthcheck run ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623. -- Subject: Unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service has finished starting up. -- -- The start-up result is done. Jul 31 21:25:03 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting quadlet-demo.service... -- Subject: Unit quadlet-demo.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo.service has begun starting up. Jul 31 21:25:03 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[59942]: Pods stopped: Jul 31 21:25:03 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[59942]: Pods removed: Jul 31 21:25:03 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[59942]: Secrets removed: Jul 31 21:25:03 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[59942]: Volumes removed: Jul 31 21:25:03 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service has successfully entered the 'dead' state. Jul 31 21:25:03 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-afecd033145f2222184550f42d9ce744a9bc9068b988a583143d888537a66895-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-afecd033145f2222184550f42d9ce744a9bc9068b988a583143d888537a66895-merged.mount has successfully entered the 'dead' state. Jul 31 21:25:03 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Created slice machine.slice. -- Subject: Unit machine.slice has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit machine.slice has finished starting up. -- -- The start-up result is done. Jul 31 21:25:03 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Created slice cgroup machine-libpod_pod_693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d.slice. -- Subject: Unit machine-libpod_pod_693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d.slice has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit machine-libpod_pod_693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d.slice has finished starting up. -- -- The start-up result is done. Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container c892d334f8f60174f658904e9a12a547283cdc7de9bdcc470a039fc1108290c5. -- Subject: Unit libpod-c892d334f8f60174f658904e9a12a547283cdc7de9bdcc470a039fc1108290c5.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-c892d334f8f60174f658904e9a12a547283cdc7de9bdcc470a039fc1108290c5.scope has finished starting up. -- -- The start-up result is done. Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475504.1887] manager: (veth8e3c8f05): new Veth device (/org/freedesktop/NetworkManager/Devices/7) Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd-udevd[60082]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable. Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd-udevd[60082]: Could not generate persistent MAC address for veth8e3c8f05: No such file or directory Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): veth8e3c8f05: link is not ready Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth8e3c8f05) entered blocking state Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth8e3c8f05) entered disabled state Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: device veth8e3c8f05 entered promiscuous mode Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_UP): eth0: link is not ready Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: IPv6: ADDRCONF(NETDEV_CHANGE): veth8e3c8f05: link becomes ready Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth8e3c8f05) entered blocking state Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth8e3c8f05) entered forwarding state Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475504.2241] device (veth8e3c8f05): carrier: link connected Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56502]: read /run/containers/cni/dnsname/systemd-quadlet-demo/addnhosts - 2 addresses Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container cd265236ca471822539294e0a696dff4df2f3680f91bc9a46fd423e6c94f9160. -- Subject: Unit libpod-cd265236ca471822539294e0a696dff4df2f3680f91bc9a46fd423e6c94f9160.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-cd265236ca471822539294e0a696dff4df2f3680f91bc9a46fd423e6c94f9160.scope has finished starting up. -- -- The start-up result is done. Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 83d50c826b2da7ffd6d1dc58298744eb859025f0616feae50af33e6332caf418. -- Subject: Unit libpod-83d50c826b2da7ffd6d1dc58298744eb859025f0616feae50af33e6332caf418.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-83d50c826b2da7ffd6d1dc58298744eb859025f0616feae50af33e6332caf418.scope has finished starting up. -- -- The start-up result is done. Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started libcontainer container 523c4ea94b7bda336384681634a6ee3b6690b3c7bd8fc0539a51bc4646041061. -- Subject: Unit libpod-523c4ea94b7bda336384681634a6ee3b6690b3c7bd8fc0539a51bc4646041061.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-523c4ea94b7bda336384681634a6ee3b6690b3c7bd8fc0539a51bc4646041061.scope has finished starting up. -- -- The start-up result is done. Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[59942]: Volumes: Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[59942]: wp-pv-claim Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[59942]: Pod: Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[59942]: 693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[59942]: Containers: Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[59942]: 523c4ea94b7bda336384681634a6ee3b6690b3c7bd8fc0539a51bc4646041061 Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[59942]: 83d50c826b2da7ffd6d1dc58298744eb859025f0616feae50af33e6332caf418 Jul 31 21:25:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started quadlet-demo.service. -- Subject: Unit quadlet-demo.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo.service has finished starting up. -- -- The start-up result is done. Jul 31 21:25:05 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[60406]: ansible-command Invoked with _raw_params=ls -alrtF /etc/containers/systemd warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:05 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[60600]: ansible-command Invoked with _raw_params=podman ps -a warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:06 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[60803]: ansible-command Invoked with _raw_params=podman volume ls warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:06 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[60949]: ansible-command Invoked with _raw_params=podman pod ps --ctr-ids --ctr-names --ctr-status warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:07 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[61080]: ansible-command Invoked with _raw_params=set -euo pipefail; systemctl list-units | grep quadlet _uses_shell=True warn=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:07 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[61206]: ansible-get_url Invoked with url=https://localhost:8000 dest=/run/out mode=0600 validate_certs=False force=False http_agent=ansible-httpget use_proxy=True force_basic_auth=False sha256sum= checksum= timeout=10 follow=False unsafe_writes=False url_username=None url_password=NOT_LOGGING_PARAMETER client_cert=None client_key=None backup=None headers=None tmp_dest=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None src=None content=NOT_LOGGING_PARAMETER remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:25:08 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[61331]: ansible-command Invoked with _raw_params=cat /run/out warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:09 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[61455]: ansible-command Invoked with _raw_params=podman ps -a warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:09 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[61585]: ansible-command Invoked with _raw_params=podman pod ps --ctr-ids --ctr-names --ctr-status warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:09 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[61716]: ansible-command Invoked with _raw_params=set -euo pipefail; systemctl list-units --all | grep quadlet _uses_shell=True warn=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:10 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[61842]: ansible-command Invoked with _raw_params=ls -alrtF /etc/systemd/system warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:13 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[62091]: ansible-command Invoked with _raw_params=podman --version warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:13 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[62220]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:25:14 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[62344]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:25:16 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[62469]: ansible-dnf Invoked with name=['firewalld'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Jul 31 21:25:18 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[62592]: ansible-systemd Invoked with name=firewalld masked=False daemon_reload=False daemon_reexec=False no_block=False state=None enabled=None force=None user=None scope=None Jul 31 21:25:19 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[62719]: ansible-systemd Invoked with name=firewalld state=started enabled=True daemon_reload=False daemon_reexec=False no_block=False force=None masked=None user=None scope=None Jul 31 21:25:19 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[62846]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['8000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Jul 31 21:25:20 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[62969]: ansible-fedora.linux_system_roles.firewall_lib Invoked with port=['9000/tcp'] permanent=True runtime=True state=enabled __report_changed=True service=[] source_port=[] forward_port=[] rich_rule=[] source=[] interface=[] interface_pci_id=[] icmp_block=[] timeout=0 ipset_entries=[] protocol=[] helper_module=[] destination=[] firewalld_conf=None masquerade=None icmp_block_inversion=None target=None zone=None set_default_zone=None ipset=None ipset_type=None description=None short=None Jul 31 21:25:23 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[63482]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:25:23 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[63606]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[63731]: ansible-systemd Invoked with name=quadlet-demo.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopping quadlet-demo.service... -- Subject: Unit quadlet-demo.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo.service has begun shutting down. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: libpod-c892d334f8f60174f658904e9a12a547283cdc7de9bdcc470a039fc1108290c5.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-c892d334f8f60174f658904e9a12a547283cdc7de9bdcc470a039fc1108290c5.scope has successfully entered the 'dead' state. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: libpod-c892d334f8f60174f658904e9a12a547283cdc7de9bdcc470a039fc1108290c5.scope: Consumed 31ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-c892d334f8f60174f658904e9a12a547283cdc7de9bdcc470a039fc1108290c5.scope completed and consumed the indicated resources. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-c892d334f8f60174f658904e9a12a547283cdc7de9bdcc470a039fc1108290c5-userdata-shm.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay\x2dcontainers-c892d334f8f60174f658904e9a12a547283cdc7de9bdcc470a039fc1108290c5-userdata-shm.mount has successfully entered the 'dead' state. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-47c6bbea6926a96d39d736072970a2121905a4fdd59a1eec0d5e72f0d1f122e7-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-47c6bbea6926a96d39d736072970a2121905a4fdd59a1eec0d5e72f0d1f122e7-merged.mount has successfully entered the 'dead' state. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: libpod-cd265236ca471822539294e0a696dff4df2f3680f91bc9a46fd423e6c94f9160.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-cd265236ca471822539294e0a696dff4df2f3680f91bc9a46fd423e6c94f9160.scope has successfully entered the 'dead' state. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: libpod-cd265236ca471822539294e0a696dff4df2f3680f91bc9a46fd423e6c94f9160.scope: Consumed 27ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-cd265236ca471822539294e0a696dff4df2f3680f91bc9a46fd423e6c94f9160.scope completed and consumed the indicated resources. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: libpod-83d50c826b2da7ffd6d1dc58298744eb859025f0616feae50af33e6332caf418.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-83d50c826b2da7ffd6d1dc58298744eb859025f0616feae50af33e6332caf418.scope has successfully entered the 'dead' state. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: libpod-83d50c826b2da7ffd6d1dc58298744eb859025f0616feae50af33e6332caf418.scope: Consumed 110ms CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-83d50c826b2da7ffd6d1dc58298744eb859025f0616feae50af33e6332caf418.scope completed and consumed the indicated resources. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com conmon[60195]: conmon 83d50c826b2da7ffd6d1 : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d.slice/libpod-83d50c826b2da7ffd6d1dc58298744eb859025f0616feae50af33e6332caf418.scope/memory.events Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: libpod-523c4ea94b7bda336384681634a6ee3b6690b3c7bd8fc0539a51bc4646041061.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-523c4ea94b7bda336384681634a6ee3b6690b3c7bd8fc0539a51bc4646041061.scope has successfully entered the 'dead' state. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: libpod-523c4ea94b7bda336384681634a6ee3b6690b3c7bd8fc0539a51bc4646041061.scope: Consumed 1.109s CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-523c4ea94b7bda336384681634a6ee3b6690b3c7bd8fc0539a51bc4646041061.scope completed and consumed the indicated resources. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com conmon[60221]: conmon 523c4ea94b7bda336384 : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d.slice/libpod-523c4ea94b7bda336384681634a6ee3b6690b3c7bd8fc0539a51bc4646041061.scope/memory.events Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-614a85b12d0b3b1dfd5f2c668fae0e20140fc5bd5e24a619871052c90a912122-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-614a85b12d0b3b1dfd5f2c668fae0e20140fc5bd5e24a619871052c90a912122-merged.mount has successfully entered the 'dead' state. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com dnsmasq[56502]: read /run/containers/cni/dnsname/systemd-quadlet-demo/addnhosts - 1 addresses Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-b855eba5bd057373964ea48efe9ff807dff3b5eaa9c3354fd4d3ba3c4a0a268c-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-b855eba5bd057373964ea48efe9ff807dff3b5eaa9c3354fd4d3ba3c4a0a268c-merged.mount has successfully entered the 'dead' state. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth8e3c8f05) entered disabled state Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: device veth8e3c8f05 left promiscuous mode Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 2(veth8e3c8f05) entered disabled state Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: run-netns-netns\x2d4fbaa8b4\x2ddfc7\x2da91b\x2dcb9f\x2d1f19d3f1fa79.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-netns-netns\x2d4fbaa8b4\x2ddfc7\x2da91b\x2dcb9f\x2d1f19d3f1fa79.mount has successfully entered the 'dead' state. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Removed slice cgroup machine-libpod_pod_693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d.slice. -- Subject: Unit machine-libpod_pod_693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d.slice has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit machine-libpod_pod_693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d.slice has finished shutting down. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: machine-libpod_pod_693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d.slice: Consumed 1.248s CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit machine-libpod_pod_693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d.slice completed and consumed the indicated resources. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[63779]: Pods stopped: Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[63779]: 693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[63779]: Pods removed: Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[63779]: 693c30debda5a0905b5f775463c998eb297d6aa8eb87a3c6d63ff0755c495d2d Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[63779]: Secrets removed: Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo[63779]: Volumes removed: Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: quadlet-demo.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-demo.service has successfully entered the 'dead' state. Jul 31 21:25:25 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-demo.service. -- Subject: Unit quadlet-demo.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo.service has finished shutting down. Jul 31 21:25:26 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[64080]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.kube follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:25:26 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-08bea1ff1255d445afb9850dfdd5e34e30ec61448f97f18dee26e6338078eb36-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-08bea1ff1255d445afb9850dfdd5e34e30ec61448f97f18dee26e6338078eb36-merged.mount has successfully entered the 'dead' state. Jul 31 21:25:26 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-cd265236ca471822539294e0a696dff4df2f3680f91bc9a46fd423e6c94f9160-userdata-shm.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay\x2dcontainers-cd265236ca471822539294e0a696dff4df2f3680f91bc9a46fd423e6c94f9160-userdata-shm.mount has successfully entered the 'dead' state. Jul 31 21:25:26 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[64328]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.kube state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:25:27 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[64451]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:25:27 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:25:27 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:27 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:28 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[64604]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:29 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[64735]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:29 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[64865]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:29 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[64995]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:30 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[65125]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:31 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[65435]: ansible-service_facts Invoked Jul 31 21:25:33 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[65647]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:25:33 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[65771]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:25:34 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started /usr/bin/podman healthcheck run ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623. -- Subject: Unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service has finished starting up. -- -- The start-up result is done. Jul 31 21:25:34 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service: Main process exited, code=exited, status=125/n/a Jul 31 21:25:34 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service: Failed with result 'exit-code'. -- Subject: Unit failed -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.service has entered the 'failed' state with result 'exit-code'. Jul 31 21:25:34 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[65904]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:25:35 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[66152]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:25:36 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[66275]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:25:36 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:25:36 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:36 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:36 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[66428]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:37 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[66558]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:37 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[66688]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:38 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[66818]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:38 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[66948]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:39 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[67258]: ansible-service_facts Invoked Jul 31 21:25:41 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[67470]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:25:41 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[67594]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:25:42 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[67719]: ansible-stat Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:25:43 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[67967]: ansible-file Invoked with path=/etc/containers/systemd/envoy-proxy-configmap.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:25:44 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[68090]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:25:44 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:25:44 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:44 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:44 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[68243]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:45 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[68373]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:45 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[68503]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:45 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[68632]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:46 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[68762]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:47 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[69072]: ansible-service_facts Invoked Jul 31 21:25:49 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[69284]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:25:49 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[69408]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:25:50 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[69533]: ansible-systemd Invoked with name=quadlet-demo-mysql.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:25:50 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:25:50 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:50 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:50 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopping quadlet-demo-mysql.service... -- Subject: Unit quadlet-demo-mysql.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql.service has begun shutting down. Jul 31 21:25:52 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.timer: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.timer has successfully entered the 'dead' state. Jul 31 21:25:52 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped /usr/bin/podman healthcheck run ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623. -- Subject: Unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.timer has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623.timer has finished shutting down. Jul 31 21:25:52 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethb66cc4ad) entered disabled state Jul 31 21:25:52 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: device vethb66cc4ad left promiscuous mode Jul 31 21:25:52 ip-10-31-45-139.us-east-1.aws.redhat.com kernel: cni-podman2: port 1(vethb66cc4ad) entered disabled state Jul 31 21:25:53 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: run-netns-netns\x2d9abfa67b\x2d57db\x2da11c\x2d541b\x2dedd795053efd.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-netns-netns\x2d9abfa67b\x2d57db\x2da11c\x2d541b\x2dedd795053efd.mount has successfully entered the 'dead' state. Jul 31 21:25:53 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623-userdata-shm.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay\x2dcontainers-ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623-userdata-shm.mount has successfully entered the 'dead' state. Jul 31 21:25:53 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay-4b977a94912eda4d74fe0e046bff3c29d4881871b47e83745826f285bf3c711d-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-4b977a94912eda4d74fe0e046bff3c29d4881871b47e83745826f285bf3c711d-merged.mount has successfully entered the 'dead' state. Jul 31 21:25:53 ip-10-31-45-139.us-east-1.aws.redhat.com quadlet-demo-mysql[69569]: ea182df3f7eff7c985466cf537cd656e421db49137275cc20e2f48b59c608623 Jul 31 21:25:53 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:25:53 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: quadlet-demo-mysql.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-demo-mysql.service has successfully entered the 'dead' state. Jul 31 21:25:53 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-demo-mysql.service. -- Subject: Unit quadlet-demo-mysql.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql.service has finished shutting down. Jul 31 21:25:53 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[69814]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:25:54 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[70062]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.container state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:25:54 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[70185]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:25:54 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:25:54 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:54 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:25:55 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:25:55 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[70468]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:56 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:25:56 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[70599]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:56 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:25:57 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[70729]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:57 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[70859]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:57 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[70989]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:25:57 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:25:58 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[71299]: ansible-service_facts Invoked Jul 31 21:26:00 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[71510]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:26:01 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[71634]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:26:02 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[71759]: ansible-systemd Invoked with name=quadlet-demo-mysql-volume.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:26:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:26:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:26:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:26:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: quadlet-demo-mysql-volume.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-demo-mysql-volume.service has successfully entered the 'dead' state. Jul 31 21:26:02 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-demo-mysql-volume.service. -- Subject: Unit quadlet-demo-mysql-volume.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-mysql-volume.service has finished shutting down. Jul 31 21:26:02 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[71916]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:26:03 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[72164]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo-mysql.volume state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:26:04 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[72287]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:26:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:26:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:26:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:26:04 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:26:05 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[72571]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:26:05 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:26:05 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[72701]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:26:06 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[72831]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:26:06 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[72961]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:26:06 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[73091]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:26:07 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:26:07 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[73401]: ansible-service_facts Invoked Jul 31 21:26:09 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[73612]: ansible-getent Invoked with database=group key=0 fail_key=False service=None split=None Jul 31 21:26:10 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[73736]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:26:11 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[73861]: ansible-systemd Invoked with name=quadlet-demo-network.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None user=None Jul 31 21:26:11 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:26:11 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:26:11 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:26:11 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: quadlet-demo-network.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit quadlet-demo-network.service has successfully entered the 'dead' state. Jul 31 21:26:11 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Stopped quadlet-demo-network.service. -- Subject: Unit quadlet-demo-network.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit quadlet-demo-network.service has finished shutting down. Jul 31 21:26:11 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[74018]: ansible-stat Invoked with path=/etc/containers/systemd/quadlet-demo.network follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 31 21:26:12 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[74266]: ansible-file Invoked with path=/etc/containers/systemd/quadlet-demo.network state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Jul 31 21:26:13 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[74389]: ansible-systemd Invoked with daemon_reload=True scope=system daemon_reexec=False no_block=False name=None state=None enabled=None force=None masked=None user=None Jul 31 21:26:13 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Reloading. Jul 31 21:26:13 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-0.slice:4: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:26:13 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: /etc/systemd/system/user-.slice.d/override.conf:2: Failed to assign slice user.slice to unit user-0.slice, ignoring: Invalid argument Jul 31 21:26:13 ip-10-31-45-139.us-east-1.aws.redhat.com NetworkManager[670]: [1722475573.8253] device (cni-podman2): state change: activated -> unmanaged (reason 'unmanaged', sys-iface-state: 'removed') Jul 31 21:26:13 ip-10-31-45-139.us-east-1.aws.redhat.com dbus-daemon[581]: [system] Activating via systemd: service name='org.freedesktop.nm_dispatcher' unit='dbus-org.freedesktop.nm-dispatcher.service' requested by ':1.3' (uid=0 pid=670 comm="/usr/sbin/NetworkManager --no-daemon " label="system_u:system_r:NetworkManager_t:s0") Jul 31 21:26:13 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:26:13 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Starting Network Manager Script Dispatcher Service... -- Subject: Unit NetworkManager-dispatcher.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has begun starting up. Jul 31 21:26:13 ip-10-31-45-139.us-east-1.aws.redhat.com dbus-daemon[581]: [system] Successfully activated service 'org.freedesktop.nm_dispatcher' Jul 31 21:26:13 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: Started Network Manager Script Dispatcher Service. -- Subject: Unit NetworkManager-dispatcher.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has finished starting up. -- -- The start-up result is done. Jul 31 21:26:14 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[74732]: ansible-command Invoked with _raw_params=podman image prune --all -f warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:26:14 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:26:14 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[74862]: ansible-command Invoked with _raw_params=podman images -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:26:15 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[74992]: ansible-command Invoked with _raw_params=podman volume ls -n warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:26:15 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[75123]: ansible-command Invoked with _raw_params=podman ps --noheading warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:26:16 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[75253]: ansible-command Invoked with _raw_params=podman network ls -n -q warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:26:16 ip-10-31-45-139.us-east-1.aws.redhat.com systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 31 21:26:16 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[75538]: ansible-service_facts Invoked Jul 31 21:26:18 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[75748]: ansible-command Invoked with _raw_params=exec 1>&2 set -x set -o pipefail systemctl list-units --plain -l --all | grep quadlet || : systemctl list-unit-files --all | grep quadlet || : systemctl list-units --plain --failed -l --all | grep quadlet || : _uses_shell=True warn=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 31 21:26:19 ip-10-31-45-139.us-east-1.aws.redhat.com platform-python[75878]: ansible-command Invoked with _raw_params=journalctl -ex warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None PLAY RECAP ********************************************************************* managed_node1 : ok=412 changed=42 unreachable=0 failed=1 skipped=368 rescued=1 ignored=0 Wednesday 31 July 2024 21:26:19 -0400 (0:00:00.463) 0:02:33.078 ******** =============================================================================== fedora.linux_system_roles.podman : Ensure container images are present -- 17.34s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 fedora.linux_system_roles.podman : Ensure container images are present --- 7.15s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 fedora.linux_system_roles.certificate : Ensure certificate role dependencies are installed --- 2.91s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:5 fedora.linux_system_roles.podman : Stop and disable service ------------- 2.76s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 fedora.linux_system_roles.certificate : Ensure provider packages are installed --- 2.44s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:23 fedora.linux_system_roles.firewall : Install firewalld ------------------ 2.41s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 fedora.linux_system_roles.firewall : Install firewalld ------------------ 2.40s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 fedora.linux_system_roles.podman : Start service ------------------------ 1.82s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:110 fedora.linux_system_roles.podman : For testing and debugging - services --- 1.68s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 fedora.linux_system_roles.podman : For testing and debugging - services --- 1.64s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 fedora.linux_system_roles.podman : Gather the package facts ------------- 1.63s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.podman : For testing and debugging - services --- 1.63s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 fedora.linux_system_roles.podman : For testing and debugging - services --- 1.59s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 fedora.linux_system_roles.podman : For testing and debugging - services --- 1.58s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 fedora.linux_system_roles.podman : For testing and debugging - services --- 1.54s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:183 fedora.linux_system_roles.podman : Gather the package facts ------------- 1.41s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 LS ---------------------------------------------------------------------- 1.36s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:153 fedora.linux_system_roles.podman : Stop and disable service ------------- 1.19s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Check web --------------------------------------------------------------- 1.13s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_demo.yml:109 fedora.linux_system_roles.certificate : Slurp the contents of the files --- 1.08s /tmp/tmp.CFhubSqMfm/ansible_collections/fedora/linux_system_roles/roles/certificate/tasks/main.yml:152