ansible-playbook [core 2.17.12] config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.12/site-packages/ansible ansible collection location = /tmp/collections-4zP executable location = /usr/local/bin/ansible-playbook python version = 3.12.11 (main, Jun 4 2025, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-7)] (/usr/bin/python3.12) jinja version = 3.1.6 libyaml = True No config file found; using defaults running playbook inside collection fedora.linux_system_roles Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'jsonl', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_quadlet_basic.yml ********************************************** 2 plays in /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml PLAY [all] ********************************************************************* TASK [Include vault variables] ************************************************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:5 Tuesday 08 July 2025 08:54:18 -0400 (0:00:00.026) 0:00:00.027 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_test_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n35383939616163653333633431363463313831383037386236646138333162396161356130303461\n3932623930643263313563336163316337643562333936360a363538636631313039343233383732\n38666530383538656639363465313230343533386130303833336434303438333161656262346562\n3362626538613031640a663330613638366132356534363534353239616666653466353961323533\n6565\n" }, "mysql_container_root_password": { "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n61333932373230333539663035366431326163363166363036323963623131363530326231303634\n6635326161643165363366323062333334363730376631660a393566366139353861656364656661\n38653463363837336639363032646433666361646535366137303464623261313663643336306465\n6264663730656337310a343962353137386238383064646533366433333437303566656433386233\n34343235326665646661623131643335313236313131353661386338343366316261643634653633\n3832313034366536616531323963333234326461353130303532\n" } }, "ansible_included_var_files": [ "/tmp/podman-AOi/tests/vars/vault-variables.yml" ], "changed": false } PLAY [Ensure that the role can manage quadlet specs] *************************** TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:9 Tuesday 08 July 2025 08:54:19 -0400 (0:00:00.071) 0:00:00.098 ********** [WARNING]: Platform linux on host managed-node1 is using the discovered Python interpreter at /usr/bin/python3.9, but future installation of another Python interpreter could change the meaning of that path. See https://docs.ansible.com/ansible- core/2.17/reference_appendices/interpreter_discovery.html for more information. ok: [managed-node1] TASK [Test is only supported on x86_64] **************************************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:55 Tuesday 08 July 2025 08:54:20 -0400 (0:00:01.158) 0:00:01.256 ********** skipping: [managed-node1] => { "false_condition": "ansible_facts[\"architecture\"] != \"x86_64\"" } TASK [End test] **************************************************************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:62 Tuesday 08 July 2025 08:54:20 -0400 (0:00:00.015) 0:00:01.271 ********** META: end_play conditional evaluated to False, continuing play skipping: [managed-node1] => { "skip_reason": "end_play conditional evaluated to False, continuing play" } MSG: end_play TASK [Run role - do not pull images] ******************************************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:70 Tuesday 08 July 2025 08:54:20 -0400 (0:00:00.008) 0:00:01.280 ********** included: fedora.linux_system_roles.podman for managed-node1 TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Tuesday 08 July 2025 08:54:20 -0400 (0:00:00.072) 0:00:01.352 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Tuesday 08 July 2025 08:54:20 -0400 (0:00:00.035) 0:00:01.388 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Tuesday 08 July 2025 08:54:20 -0400 (0:00:00.042) 0:00:01.430 ********** ok: [managed-node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Tuesday 08 July 2025 08:54:20 -0400 (0:00:00.452) 0:00:01.883 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Tuesday 08 July 2025 08:54:20 -0400 (0:00:00.029) 0:00:01.912 ********** ok: [managed-node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Tuesday 08 July 2025 08:54:21 -0400 (0:00:00.355) 0:00:02.268 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_is_transactional": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Tuesday 08 July 2025 08:54:21 -0400 (0:00:00.023) 0:00:02.292 ********** ok: [managed-node1] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed-node1] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node1] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node1] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Tuesday 08 July 2025 08:54:21 -0400 (0:00:00.050) 0:00:02.342 ********** ok: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Tuesday 08 July 2025 08:54:22 -0400 (0:00:01.044) 0:00:03.386 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Tuesday 08 July 2025 08:54:22 -0400 (0:00:00.048) 0:00:03.434 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages)) | list | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Tuesday 08 July 2025 08:54:22 -0400 (0:00:00.063) 0:00:03.497 ********** skipping: [managed-node1] => { "false_condition": "__podman_is_transactional | d(false)" } TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Tuesday 08 July 2025 08:54:22 -0400 (0:00:00.071) 0:00:03.569 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Tuesday 08 July 2025 08:54:22 -0400 (0:00:00.067) 0:00:03.636 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Tuesday 08 July 2025 08:54:22 -0400 (0:00:00.061) 0:00:03.698 ********** ok: [managed-node1] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.025551", "end": "2025-07-08 08:54:23.075656", "rc": 0, "start": "2025-07-08 08:54:23.050105" } STDOUT: podman version 5.5.1 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Tuesday 08 July 2025 08:54:23 -0400 (0:00:00.485) 0:00:04.184 ********** ok: [managed-node1] => { "ansible_facts": { "podman_version": "5.5.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Tuesday 08 July 2025 08:54:23 -0400 (0:00:00.044) 0:00:04.228 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Tuesday 08 July 2025 08:54:23 -0400 (0:00:00.054) 0:00:04.283 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Tuesday 08 July 2025 08:54:23 -0400 (0:00:00.065) 0:00:04.349 ********** META: end_host conditional evaluated to False, continuing execution for managed-node1 skipping: [managed-node1] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node1" } MSG: end_host conditional evaluated to false, continuing execution for managed-node1 TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Tuesday 08 July 2025 08:54:23 -0400 (0:00:00.072) 0:00:04.422 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__has_type_pod or __has_pod_file_ext or __has_pod_file_src_ext or __has_pod_template_src_ext or __has_pod_template_src_ext_j2", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:96 Tuesday 08 July 2025 08:54:23 -0400 (0:00:00.085) 0:00:04.507 ********** META: end_host conditional evaluated to False, continuing execution for managed-node1 skipping: [managed-node1] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node1" } MSG: end_host conditional evaluated to false, continuing execution for managed-node1 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:109 Tuesday 08 July 2025 08:54:23 -0400 (0:00:00.046) 0:00:04.554 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 08 July 2025 08:54:23 -0400 (0:00:00.064) 0:00:04.618 ********** ok: [managed-node1] => { "ansible_facts": { "getent_passwd": { "root": [ "x", "0", "0", "root", "/root", "/bin/bash" ] } }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 08 July 2025 08:54:24 -0400 (0:00:00.509) 0:00:05.127 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 08 July 2025 08:54:24 -0400 (0:00:00.062) 0:00:05.190 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Tuesday 08 July 2025 08:54:24 -0400 (0:00:00.070) 0:00:05.260 ********** ok: [managed-node1] => { "changed": false, "stat": { "atime": 1751979003.1280832, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "8bedde2dbca15219e1a3b95a68a8c0d26a92ba62", "ctime": 1751978976.6998076, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 665568, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1748273472.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "4278445899", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Tuesday 08 July 2025 08:54:24 -0400 (0:00:00.428) 0:00:05.689 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Tuesday 08 July 2025 08:54:24 -0400 (0:00:00.039) 0:00:05.728 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Tuesday 08 July 2025 08:54:24 -0400 (0:00:00.036) 0:00:05.765 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Tuesday 08 July 2025 08:54:24 -0400 (0:00:00.040) 0:00:05.805 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Tuesday 08 July 2025 08:54:24 -0400 (0:00:00.040) 0:00:05.846 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Tuesday 08 July 2025 08:54:24 -0400 (0:00:00.038) 0:00:05.884 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Tuesday 08 July 2025 08:54:24 -0400 (0:00:00.032) 0:00:05.917 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Tuesday 08 July 2025 08:54:24 -0400 (0:00:00.033) 0:00:05.950 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:115 Tuesday 08 July 2025 08:54:24 -0400 (0:00:00.032) 0:00:05.983 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_parent_mode": "0755", "__podman_parent_path": "/etc/containers", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:126 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.070) 0:00:06.053 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.099) 0:00:06.153 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.100) 0:00:06.253 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:129 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.052) 0:00:06.306 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.105) 0:00:06.411 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.071) 0:00:06.483 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:132 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.052) 0:00:06.535 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:7 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.108) 0:00:06.644 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:15 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.057) 0:00:06.701 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:135 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.054) 0:00:06.756 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:8 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.113) 0:00:06.869 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:16 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.055) 0:00:06.925 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:21 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.053) 0:00:06.978 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:27 Tuesday 08 July 2025 08:54:25 -0400 (0:00:00.054) 0:00:07.033 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:141 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.053) 0:00:07.087 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_firewall | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage selinux for specified ports] ************************************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:148 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.052) 0:00:07.139 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:155 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.054) 0:00:07.194 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:159 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.052) 0:00:07.247 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:168 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.097) 0:00:07.344 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:177 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.032) 0:00:07.377 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:184 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.032) 0:00:07.409 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:191 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.033) 0:00:07.443 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node1 => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.101) 0:00:07.544 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": { "Container": { "ContainerName": "nopull", "Image": "quay.io/libpod/testimage:20210610" }, "Install": { "WantedBy": "default.target" } }, "__podman_quadlet_str": "", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.041) 0:00:07.586 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": false, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.040) 0:00:07.627 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_quadlet_spec | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.032) 0:00:07.659 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_quadlet_name": "nopull", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.060) 0:00:07.720 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.104) 0:00:07.824 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.057) 0:00:07.882 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.059) 0:00:07.942 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Tuesday 08 July 2025 08:54:26 -0400 (0:00:00.072) 0:00:08.014 ********** ok: [managed-node1] => { "changed": false, "stat": { "atime": 1751979003.1280832, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "8bedde2dbca15219e1a3b95a68a8c0d26a92ba62", "ctime": 1751978976.6998076, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 665568, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1748273472.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "4278445899", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.422) 0:00:08.437 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.041) 0:00:08.479 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.049) 0:00:08.528 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.040) 0:00:08.569 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.076) 0:00:08.645 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.033) 0:00:08.678 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.034) 0:00:08.713 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.033) 0:00:08.747 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.038) 0:00:08.786 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_activate_systemd_unit": false, "__podman_images_found": [ "quay.io/libpod/testimage:20210610" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "nopull.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.073) 0:00:08.859 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.041) 0:00:08.900 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:88 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.043) 0:00:08.943 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_images": [ "quay.io/libpod/testimage:20210610" ], "__podman_quadlet_file": "/etc/containers/systemd/nopull.container", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:106 Tuesday 08 July 2025 08:54:27 -0400 (0:00:00.089) 0:00:09.033 ********** ok: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:113 Tuesday 08 July 2025 08:54:28 -0400 (0:00:00.038) 0:00:09.071 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:117 Tuesday 08 July 2025 08:54:28 -0400 (0:00:00.036) 0:00:09.108 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Tuesday 08 July 2025 08:54:28 -0400 (0:00:00.078) 0:00:09.187 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 08 July 2025 08:54:28 -0400 (0:00:00.080) 0:00:09.268 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 08 July 2025 08:54:28 -0400 (0:00:00.051) 0:00:09.320 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 08 July 2025 08:54:28 -0400 (0:00:00.039) 0:00:09.359 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Tuesday 08 July 2025 08:54:28 -0400 (0:00:00.034) 0:00:09.393 ********** skipping: [managed-node1] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Tuesday 08 July 2025 08:54:28 -0400 (0:00:00.037) 0:00:09.430 ********** skipping: [managed-node1] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Tuesday 08 July 2025 08:54:28 -0400 (0:00:00.046) 0:00:09.477 ********** ok: [managed-node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 43, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:50 Tuesday 08 July 2025 08:54:29 -0400 (0:00:00.564) 0:00:10.041 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_quadlet_file_src | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:62 Tuesday 08 July 2025 08:54:29 -0400 (0:00:00.034) 0:00:10.075 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_quadlet_str | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:75 Tuesday 08 July 2025 08:54:29 -0400 (0:00:00.036) 0:00:10.112 ********** changed: [managed-node1] => { "changed": true, "checksum": "670d64fc68a9768edb20cad26df2acc703542d85", "dest": "/etc/containers/systemd/nopull.container", "gid": 0, "group": "root", "md5sum": "cedb6667f6cd1b033fe06e2810fe6b19", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 151, "src": "/root/.ansible/tmp/ansible-tmp-1751979269.1189396-16954-11122430624534/.source.container", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:87 Tuesday 08 July 2025 08:54:29 -0400 (0:00:00.855) 0:00:10.967 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_activate_systemd_unit | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:115 Tuesday 08 July 2025 08:54:29 -0400 (0:00:00.046) 0:00:11.014 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_activate_systemd_unit | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:131 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.044) 0:00:11.058 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_activate_systemd_unit | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:198 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.046) 0:00:11.105 ********** skipping: [managed-node1] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle credential files - absent] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:204 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.038) 0:00:11.143 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ******** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:213 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.034) 0:00:11.177 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [Verify image not pulled] ************************************************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:86 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.047) 0:00:11.225 ********** ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [Run role - try to pull bogus image] ************************************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:90 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.035) 0:00:11.261 ********** included: fedora.linux_system_roles.podman for managed-node1 TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.105) 0:00:11.366 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.093) 0:00:11.460 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.063) 0:00:11.523 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.058) 0:00:11.582 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.043) 0:00:11.626 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.087) 0:00:11.713 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.037) 0:00:11.751 ********** ok: [managed-node1] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed-node1] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node1] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node1] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Tuesday 08 July 2025 08:54:30 -0400 (0:00:00.067) 0:00:11.819 ********** ok: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Tuesday 08 July 2025 08:54:31 -0400 (0:00:00.820) 0:00:12.639 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Tuesday 08 July 2025 08:54:31 -0400 (0:00:00.035) 0:00:12.675 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages)) | list | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Tuesday 08 July 2025 08:54:31 -0400 (0:00:00.064) 0:00:12.739 ********** skipping: [managed-node1] => { "false_condition": "__podman_is_transactional | d(false)" } TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Tuesday 08 July 2025 08:54:31 -0400 (0:00:00.054) 0:00:12.794 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Tuesday 08 July 2025 08:54:31 -0400 (0:00:00.051) 0:00:12.846 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Tuesday 08 July 2025 08:54:31 -0400 (0:00:00.053) 0:00:12.900 ********** ok: [managed-node1] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.025880", "end": "2025-07-08 08:54:32.217260", "rc": 0, "start": "2025-07-08 08:54:32.191380" } STDOUT: podman version 5.5.1 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Tuesday 08 July 2025 08:54:32 -0400 (0:00:00.429) 0:00:13.329 ********** ok: [managed-node1] => { "ansible_facts": { "podman_version": "5.5.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Tuesday 08 July 2025 08:54:32 -0400 (0:00:00.043) 0:00:13.372 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Tuesday 08 July 2025 08:54:32 -0400 (0:00:00.036) 0:00:13.409 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Tuesday 08 July 2025 08:54:32 -0400 (0:00:00.035) 0:00:13.444 ********** META: end_host conditional evaluated to False, continuing execution for managed-node1 skipping: [managed-node1] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node1" } MSG: end_host conditional evaluated to false, continuing execution for managed-node1 TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Tuesday 08 July 2025 08:54:32 -0400 (0:00:00.037) 0:00:13.482 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__has_type_pod or __has_pod_file_ext or __has_pod_file_src_ext or __has_pod_template_src_ext or __has_pod_template_src_ext_j2", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:96 Tuesday 08 July 2025 08:54:32 -0400 (0:00:00.044) 0:00:13.526 ********** META: end_host conditional evaluated to False, continuing execution for managed-node1 skipping: [managed-node1] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node1" } MSG: end_host conditional evaluated to false, continuing execution for managed-node1 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:109 Tuesday 08 July 2025 08:54:32 -0400 (0:00:00.084) 0:00:13.611 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 08 July 2025 08:54:32 -0400 (0:00:00.077) 0:00:13.689 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 08 July 2025 08:54:32 -0400 (0:00:00.050) 0:00:13.739 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 08 July 2025 08:54:32 -0400 (0:00:00.043) 0:00:13.782 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Tuesday 08 July 2025 08:54:32 -0400 (0:00:00.051) 0:00:13.834 ********** ok: [managed-node1] => { "changed": false, "stat": { "atime": 1751979003.1280832, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "8bedde2dbca15219e1a3b95a68a8c0d26a92ba62", "ctime": 1751978976.6998076, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 665568, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1748273472.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "4278445899", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.377) 0:00:14.211 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.043) 0:00:14.254 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.038) 0:00:14.292 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.035) 0:00:14.327 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.034) 0:00:14.362 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.034) 0:00:14.397 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.033) 0:00:14.430 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.036) 0:00:14.467 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:115 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.043) 0:00:14.511 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_parent_mode": "0755", "__podman_parent_path": "/etc/containers", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:126 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.068) 0:00:14.580 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.121) 0:00:14.701 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.044) 0:00:14.746 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:129 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.036) 0:00:14.782 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.120) 0:00:14.903 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.032) 0:00:14.936 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:132 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.033) 0:00:14.970 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:7 Tuesday 08 July 2025 08:54:33 -0400 (0:00:00.063) 0:00:15.034 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:15 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.034) 0:00:15.068 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:135 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.041) 0:00:15.110 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:8 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.079) 0:00:15.189 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:16 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.038) 0:00:15.228 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:21 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.040) 0:00:15.268 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:27 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.038) 0:00:15.306 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:141 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.033) 0:00:15.340 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_firewall | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage selinux for specified ports] ************************************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:148 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.032) 0:00:15.373 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:155 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.033) 0:00:15.407 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:159 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.032) 0:00:15.439 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:168 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.034) 0:00:15.474 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:177 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.039) 0:00:15.514 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:184 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.083) 0:00:15.597 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:191 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.035) 0:00:15.632 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node1 => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.096) 0:00:15.729 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": { "Container": { "ContainerName": "bogus", "Image": "this_is_a_bogus_image" }, "Install": { "WantedBy": "default.target" } }, "__podman_quadlet_str": "", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.042) 0:00:15.771 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": true, "__podman_pull_image": true, "__podman_state": "created", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.042) 0:00:15.814 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_quadlet_spec | length == 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.035) 0:00:15.850 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_quadlet_name": "bogus", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.066) 0:00:15.917 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 08 July 2025 08:54:34 -0400 (0:00:00.086) 0:00:16.003 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.045) 0:00:16.048 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.044) 0:00:16.093 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.050) 0:00:16.144 ********** ok: [managed-node1] => { "changed": false, "stat": { "atime": 1751979003.1280832, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "8bedde2dbca15219e1a3b95a68a8c0d26a92ba62", "ctime": 1751978976.6998076, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 665568, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1748273472.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "4278445899", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.386) 0:00:16.530 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.039) 0:00:16.570 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.034) 0:00:16.604 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.035) 0:00:16.639 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.034) 0:00:16.674 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.034) 0:00:16.709 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.085) 0:00:16.794 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.040) 0:00:16.834 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.044) 0:00:16.879 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_activate_systemd_unit": false, "__podman_images_found": [ "this_is_a_bogus_image" ], "__podman_kube_yamls_raw": "", "__podman_service_name": "bogus.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.067) 0:00:16.947 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.048) 0:00:16.995 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_kube_yamls_raw | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:88 Tuesday 08 July 2025 08:54:35 -0400 (0:00:00.040) 0:00:17.036 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_images": [ "this_is_a_bogus_image" ], "__podman_quadlet_file": "/etc/containers/systemd/bogus.container", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:106 Tuesday 08 July 2025 08:54:36 -0400 (0:00:00.077) 0:00:17.113 ********** ok: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:113 Tuesday 08 July 2025 08:54:36 -0400 (0:00:00.040) 0:00:17.153 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_state == \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create and update quadlets] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:117 Tuesday 08 July 2025 08:54:36 -0400 (0:00:00.032) 0:00:17.186 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Manage linger] ************************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:2 Tuesday 08 July 2025 08:54:36 -0400 (0:00:00.072) 0:00:17.258 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Enable linger if needed] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:12 Tuesday 08 July 2025 08:54:36 -0400 (0:00:00.084) 0:00:17.343 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user as not yet needing to cancel linger] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:18 Tuesday 08 July 2025 08:54:36 -0400 (0:00:00.055) 0:00:17.399 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Mark user for possible linger cancel] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/manage_linger.yml:22 Tuesday 08 July 2025 08:54:36 -0400 (0:00:00.054) 0:00:17.453 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Create host directories] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:7 Tuesday 08 July 2025 08:54:36 -0400 (0:00:00.054) 0:00:17.507 ********** skipping: [managed-node1] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Ensure container images are present] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 Tuesday 08 July 2025 08:54:36 -0400 (0:00:00.048) 0:00:17.556 ********** ok: [managed-node1] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } ok: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Ensure the quadlet directory is present] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Tuesday 08 July 2025 08:54:37 -0400 (0:00:00.701) 0:00:18.258 ********** ok: [managed-node1] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/containers/systemd", "secontext": "system_u:object_r:etc_t:s0", "size": 67, "state": "directory", "uid": 0 } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is copied] ******** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:50 Tuesday 08 July 2025 08:54:37 -0400 (0:00:00.391) 0:00:18.649 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_quadlet_file_src | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file content is present] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:62 Tuesday 08 July 2025 08:54:37 -0400 (0:00:00.034) 0:00:18.684 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_quadlet_str | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure quadlet file is present] ******* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:75 Tuesday 08 July 2025 08:54:37 -0400 (0:00:00.078) 0:00:18.763 ********** changed: [managed-node1] => { "changed": true, "checksum": "1d087e679d135214e8ac9ccaf33b2222916efb7f", "dest": "/etc/containers/systemd/bogus.container", "gid": 0, "group": "root", "md5sum": "97480a9a73734d9f8007d2c06e7fed1f", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 138, "src": "/root/.ansible/tmp/ansible-tmp-1751979277.771543-17272-2966124504745/.source.container", "state": "file", "uid": 0 } TASK [fedora.linux_system_roles.podman : Reload systemctl] ********************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:87 Tuesday 08 July 2025 08:54:38 -0400 (0:00:00.740) 0:00:19.504 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_activate_systemd_unit | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Start service] ************************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:115 Tuesday 08 July 2025 08:54:38 -0400 (0:00:00.036) 0:00:19.540 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_activate_systemd_unit | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Restart service] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:131 Tuesday 08 July 2025 08:54:38 -0400 (0:00:00.038) 0:00:19.579 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_activate_systemd_unit | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Cancel linger] ************************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:198 Tuesday 08 July 2025 08:54:38 -0400 (0:00:00.040) 0:00:19.620 ********** skipping: [managed-node1] => { "changed": false, "skipped_reason": "No items in the list" } TASK [fedora.linux_system_roles.podman : Handle credential files - absent] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:204 Tuesday 08 July 2025 08:54:38 -0400 (0:00:00.029) 0:00:19.649 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - absent] ******** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:213 Tuesday 08 July 2025 08:54:38 -0400 (0:00:00.035) 0:00:19.685 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [Verify image not pulled and no error] ************************************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:106 Tuesday 08 July 2025 08:54:38 -0400 (0:00:00.064) 0:00:19.749 ********** ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [Cleanup] ***************************************************************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:113 Tuesday 08 July 2025 08:54:38 -0400 (0:00:00.060) 0:00:19.809 ********** included: fedora.linux_system_roles.podman for managed-node1 => (item=nopull) included: fedora.linux_system_roles.podman for managed-node1 => (item=bogus) TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Tuesday 08 July 2025 08:54:39 -0400 (0:00:00.296) 0:00:20.106 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Tuesday 08 July 2025 08:54:39 -0400 (0:00:00.090) 0:00:20.197 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Tuesday 08 July 2025 08:54:39 -0400 (0:00:00.065) 0:00:20.262 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Tuesday 08 July 2025 08:54:39 -0400 (0:00:00.052) 0:00:20.315 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Tuesday 08 July 2025 08:54:39 -0400 (0:00:00.054) 0:00:20.370 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Tuesday 08 July 2025 08:54:39 -0400 (0:00:00.054) 0:00:20.424 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Tuesday 08 July 2025 08:54:39 -0400 (0:00:00.114) 0:00:20.539 ********** [WARNING]: TASK: fedora.linux_system_roles.podman : Set platform/version specific variables: The loop variable 'item' is already in use. You should set the `loop_var` value in the `loop_control` option for the task to something else to avoid variable collisions and unexpected behavior. ok: [managed-node1] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed-node1] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node1] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node1] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Tuesday 08 July 2025 08:54:39 -0400 (0:00:00.109) 0:00:20.649 ********** ok: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Tuesday 08 July 2025 08:54:40 -0400 (0:00:00.922) 0:00:21.572 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Tuesday 08 July 2025 08:54:40 -0400 (0:00:00.056) 0:00:21.628 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages)) | list | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Tuesday 08 July 2025 08:54:40 -0400 (0:00:00.066) 0:00:21.695 ********** skipping: [managed-node1] => { "false_condition": "__podman_is_transactional | d(false)" } TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Tuesday 08 July 2025 08:54:40 -0400 (0:00:00.055) 0:00:21.750 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Tuesday 08 July 2025 08:54:40 -0400 (0:00:00.058) 0:00:21.808 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Tuesday 08 July 2025 08:54:40 -0400 (0:00:00.054) 0:00:21.863 ********** ok: [managed-node1] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.026783", "end": "2025-07-08 08:54:41.192033", "rc": 0, "start": "2025-07-08 08:54:41.165250" } STDOUT: podman version 5.5.1 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Tuesday 08 July 2025 08:54:41 -0400 (0:00:00.454) 0:00:22.318 ********** ok: [managed-node1] => { "ansible_facts": { "podman_version": "5.5.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Tuesday 08 July 2025 08:54:41 -0400 (0:00:00.061) 0:00:22.380 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Tuesday 08 July 2025 08:54:41 -0400 (0:00:00.054) 0:00:22.434 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Tuesday 08 July 2025 08:54:41 -0400 (0:00:00.056) 0:00:22.491 ********** META: end_host conditional evaluated to False, continuing execution for managed-node1 skipping: [managed-node1] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node1" } MSG: end_host conditional evaluated to false, continuing execution for managed-node1 TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Tuesday 08 July 2025 08:54:41 -0400 (0:00:00.064) 0:00:22.555 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__has_type_pod or __has_pod_file_ext or __has_pod_file_src_ext or __has_pod_template_src_ext or __has_pod_template_src_ext_j2", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:96 Tuesday 08 July 2025 08:54:41 -0400 (0:00:00.075) 0:00:22.630 ********** META: end_host conditional evaluated to False, continuing execution for managed-node1 skipping: [managed-node1] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node1" } MSG: end_host conditional evaluated to false, continuing execution for managed-node1 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:109 Tuesday 08 July 2025 08:54:41 -0400 (0:00:00.081) 0:00:22.712 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 08 July 2025 08:54:41 -0400 (0:00:00.165) 0:00:22.877 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 08 July 2025 08:54:41 -0400 (0:00:00.070) 0:00:22.948 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 08 July 2025 08:54:41 -0400 (0:00:00.065) 0:00:23.013 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Tuesday 08 July 2025 08:54:42 -0400 (0:00:00.074) 0:00:23.088 ********** ok: [managed-node1] => { "changed": false, "stat": { "atime": 1751979003.1280832, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "8bedde2dbca15219e1a3b95a68a8c0d26a92ba62", "ctime": 1751978976.6998076, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 665568, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1748273472.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "4278445899", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Tuesday 08 July 2025 08:54:42 -0400 (0:00:00.422) 0:00:23.511 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Tuesday 08 July 2025 08:54:42 -0400 (0:00:00.058) 0:00:23.569 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Tuesday 08 July 2025 08:54:42 -0400 (0:00:00.057) 0:00:23.626 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Tuesday 08 July 2025 08:54:42 -0400 (0:00:00.059) 0:00:23.686 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Tuesday 08 July 2025 08:54:42 -0400 (0:00:00.056) 0:00:23.743 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Tuesday 08 July 2025 08:54:42 -0400 (0:00:00.059) 0:00:23.802 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Tuesday 08 July 2025 08:54:42 -0400 (0:00:00.061) 0:00:23.864 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Tuesday 08 July 2025 08:54:42 -0400 (0:00:00.056) 0:00:23.921 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set config file paths] **************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:115 Tuesday 08 July 2025 08:54:42 -0400 (0:00:00.060) 0:00:23.981 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_container_conf_file": "/etc/containers/containers.conf.d/50-systemroles.conf", "__podman_parent_mode": "0755", "__podman_parent_path": "/etc/containers", "__podman_policy_json_file": "/etc/containers/policy.json", "__podman_registries_conf_file": "/etc/containers/registries.conf.d/50-systemroles.conf", "__podman_storage_conf_file": "/etc/containers/storage.conf" }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle container.conf.d] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:126 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.074) 0:00:24.055 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure containers.d exists] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:5 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.107) 0:00:24.163 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update container config file] ********* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_container_conf_d.yml:13 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.057) 0:00:24.221 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_containers_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle registries.conf.d] ************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:129 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.057) 0:00:24.278 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure registries.d exists] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:5 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.164) 0:00:24.442 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update registries config file] ******** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_registries_conf_d.yml:13 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.038) 0:00:24.481 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_registries_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle storage.conf] ****************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:132 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.042) 0:00:24.523 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure storage.conf parent dir exists] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:7 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.077) 0:00:24.601 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Update storage config file] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_storage_conf.yml:15 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.032) 0:00:24.634 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_storage_conf | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Handle policy.json] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:135 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.033) 0:00:24.668 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure policy.json parent dir exists] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:8 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.067) 0:00:24.735 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stat the policy.json file] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:16 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.039) 0:00:24.775 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get the existing policy.json] ********* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:21 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.053) 0:00:24.829 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Write new policy.json file] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_policy_json.yml:27 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.055) 0:00:24.884 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_policy_json | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage firewall for specified ports] ************************************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:141 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.054) 0:00:24.939 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_firewall | length > 0", "skip_reason": "Conditional result was False" } TASK [Manage selinux for specified ports] ************************************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:148 Tuesday 08 July 2025 08:54:43 -0400 (0:00:00.055) 0:00:24.995 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_selinux_ports | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Keep track of users that need to cancel linger] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:155 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.064) 0:00:25.059 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_cancel_user_linger": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Handle certs.d files - present] ******* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:159 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.045) 0:00:25.105 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle credential files - present] **** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:168 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.034) 0:00:25.140 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle secrets] *********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:177 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.038) 0:00:25.178 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Kubernetes specifications] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:184 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.037) 0:00:25.216 ********** skipping: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Handle Quadlet specifications] ******** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:191 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.034) 0:00:25.250 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml for managed-node1 => (item=(censored due to no_log)) TASK [fedora.linux_system_roles.podman : Set per-container variables part 0] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:14 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.124) 0:00:25.374 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_quadlet_file_src": "", "__podman_quadlet_spec": {}, "__podman_quadlet_str": "", "__podman_quadlet_template_src": "" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 1] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:25 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.047) 0:00:25.422 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_continue_if_pull_fails": false, "__podman_pull_image": true, "__podman_state": "absent", "__podman_systemd_unit_scope": "", "__podman_user": "root" }, "changed": false } TASK [fedora.linux_system_roles.podman : Fail if no quadlet spec is given] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:35 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.060) 0:00:25.483 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 2] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:48 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.059) 0:00:25.542 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_quadlet_name": "nopull", "__podman_quadlet_type": "container", "__podman_rootless": false }, "changed": false } TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:57 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.053) 0:00:25.596 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.081) 0:00:25.678 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "'getent_passwd' not in ansible_facts or __podman_user not in ansible_facts['getent_passwd']", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.042) 0:00:25.720 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not ansible_facts[\"getent_passwd\"][__podman_user]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set group for podman user] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:16 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.037) 0:00:25.758 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_group": "0" }, "changed": false } TASK [fedora.linux_system_roles.podman : See if getsubids exists] ************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 Tuesday 08 July 2025 08:54:44 -0400 (0:00:00.047) 0:00:25.805 ********** ok: [managed-node1] => { "changed": false, "stat": { "atime": 1751979003.1280832, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 32, "charset": "binary", "checksum": "8bedde2dbca15219e1a3b95a68a8c0d26a92ba62", "ctime": 1751978976.6998076, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 665568, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/x-pie-executable", "mode": "0755", "mtime": 1748273472.0, "nlink": 1, "path": "/usr/bin/getsubids", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 15496, "uid": 0, "version": "4278445899", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subuids] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:42 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.389) 0:00:26.194 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check with getsubids for user subgids] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:47 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.035) 0:00:26.229 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:52 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.036) 0:00:26.266 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_user not in [\"root\", \"0\"]", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subuid file] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:65 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.043) 0:00:26.309 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get subgid file] ********************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:70 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.057) 0:00:26.367 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set user subuid and subgid info] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:75 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.060) 0:00:26.427 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subuid file] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:85 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.058) 0:00:26.486 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if user not in subgid file] ****** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:92 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.059) 0:00:26.545 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_stat_getsubids.stat.exists", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 3] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:62 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.119) 0:00:26.665 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_activate_systemd_unit": true, "__podman_images_found": [], "__podman_kube_yamls_raw": "", "__podman_service_name": "nopull.service", "__podman_systemd_scope": "system", "__podman_user_home_dir": "/root", "__podman_xdg_runtime_dir": "/run/user/0" }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 4] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:73 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.092) 0:00:26.757 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_quadlet_path": "/etc/containers/systemd" }, "changed": false } TASK [fedora.linux_system_roles.podman : Get kube yaml contents] *************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:77 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.062) 0:00:26.820 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_state != \"absent\"", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set per-container variables part 5] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:88 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.055) 0:00:26.875 ********** ok: [managed-node1] => { "ansible_facts": { "__podman_images": [], "__podman_quadlet_file": "/etc/containers/systemd/nopull.container", "__podman_volumes": [] }, "changed": false } TASK [fedora.linux_system_roles.podman : Set per-container variables part 6] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:106 Tuesday 08 July 2025 08:54:45 -0400 (0:00:00.130) 0:00:27.006 ********** ok: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Cleanup quadlets] ********************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_quadlet_spec.yml:113 Tuesday 08 July 2025 08:54:46 -0400 (0:00:00.066) 0:00:27.072 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Stat XDG_RUNTIME_DIR] ***************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:4 Tuesday 08 July 2025 08:54:46 -0400 (0:00:00.148) 0:00:27.221 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_rootless | bool", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Stop and disable service] ************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 Tuesday 08 July 2025 08:54:46 -0400 (0:00:00.057) 0:00:27.278 ********** ok: [managed-node1] => { "changed": false, "failed_when_result": false } MSG: Could not find the requested service nopull.service: host TASK [fedora.linux_system_roles.podman : See if quadlet file exists] *********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:34 Tuesday 08 July 2025 08:54:47 -0400 (0:00:00.803) 0:00:28.082 ********** ok: [managed-node1] => { "changed": false, "stat": { "atime": 1751979269.8420296, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "670d64fc68a9768edb20cad26df2acc703542d85", "ctime": 1751979269.8460295, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 226492622, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1751979269.4400246, "nlink": 1, "path": "/etc/containers/systemd/nopull.container", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 151, "uid": 0, "version": "1808348170", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:39 Tuesday 08 July 2025 08:54:47 -0400 (0:00:00.386) 0:00:28.468 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Slurp quadlet file] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 Tuesday 08 July 2025 08:54:47 -0400 (0:00:00.070) 0:00:28.539 ********** ok: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Parse quadlet file] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12 Tuesday 08 July 2025 08:54:47 -0400 (0:00:00.455) 0:00:28.994 ********** fatal: [managed-node1]: FAILED! => {} MSG: template error while templating string: Could not load "podman_from_ini": 'podman_from_ini'. String: {{ __podman_quadlet_raw.content | b64decode | podman_from_ini }}. Could not load "podman_from_ini": 'podman_from_ini' TASK [Debug3] ****************************************************************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:270 Tuesday 08 July 2025 08:54:47 -0400 (0:00:00.040) 0:00:29.035 ********** fatal: [managed-node1]: FAILED! => { "changed": false, "cmd": "set -x\nset -o pipefail\nexec 1>&2\n#podman volume rm --all\n#podman network prune -f\npodman volume ls\npodman network ls\npodman secret ls\npodman container ls\npodman pod ls\npodman images\nsystemctl list-units | grep quadlet\n", "delta": "0:00:00.191455", "end": "2025-07-08 08:54:48.480918", "rc": 1, "start": "2025-07-08 08:54:48.289463" } STDERR: + set -o pipefail + exec + podman volume ls DRIVER VOLUME NAME + podman network ls NETWORK ID NAME DRIVER 2f259bab93aa podman bridge 755dd383fce6 podman-default-kube-network bridge + podman secret ls ID NAME DRIVER CREATED UPDATED + podman container ls CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES + podman pod ls POD ID NAME STATUS CREATED INFRA ID # OF CONTAINERS + podman images REPOSITORY TAG IMAGE ID CREATED SIZE quay.io/libpod/registry 2.8.2 0030ba3d620c 23 months ago 24.6 MB localhost:5000/libpod/testimage 20210610 9f9ec7f2fdef 4 years ago 7.99 MB quay.io/libpod/testimage 20210610 9f9ec7f2fdef 4 years ago 7.99 MB + systemctl list-units + grep quadlet MSG: non-zero return code TASK [Cleanup user] ************************************************************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:299 Tuesday 08 July 2025 08:54:48 -0400 (0:00:00.551) 0:00:29.586 ********** included: fedora.linux_system_roles.podman for managed-node1 TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:3 Tuesday 08 July 2025 08:54:48 -0400 (0:00:00.078) 0:00:29.665 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Ensure ansible_facts used by role] **** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:3 Tuesday 08 July 2025 08:54:48 -0400 (0:00:00.104) 0:00:29.769 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if system is ostree] ************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Tuesday 08 July 2025 08:54:48 -0400 (0:00:00.041) 0:00:29.811 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag to indicate system is ostree] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:16 Tuesday 08 July 2025 08:54:48 -0400 (0:00:00.033) 0:00:29.844 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_is_ostree is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:23 Tuesday 08 July 2025 08:54:48 -0400 (0:00:00.034) 0:00:29.879 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set flag if transactional-update exists] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:28 Tuesday 08 July 2025 08:54:48 -0400 (0:00:00.034) 0:00:29.913 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "not __podman_is_transactional is defined", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Set platform/version specific variables] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:32 Tuesday 08 July 2025 08:54:48 -0400 (0:00:00.033) 0:00:29.947 ********** ok: [managed-node1] => (item=RedHat.yml) => { "ansible_facts": { "__podman_packages": [ "podman", "shadow-utils-subid" ] }, "ansible_included_var_files": [ "/tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/vars/RedHat.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml" } skipping: [managed-node1] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node1] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node1] => (item=CentOS_9.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS_9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Gather the package facts] ************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 Tuesday 08 July 2025 08:54:48 -0400 (0:00:00.067) 0:00:30.014 ********** ok: [managed-node1] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } TASK [fedora.linux_system_roles.podman : Enable copr if requested] ************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:10 Tuesday 08 July 2025 08:54:49 -0400 (0:00:00.811) 0:00:30.826 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_use_copr | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Ensure required packages are installed] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:14 Tuesday 08 July 2025 08:54:49 -0400 (0:00:00.033) 0:00:30.859 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "(__podman_packages | difference(ansible_facts.packages)) | list | length > 0", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:28 Tuesday 08 July 2025 08:54:49 -0400 (0:00:00.040) 0:00:30.899 ********** skipping: [managed-node1] => { "false_condition": "__podman_is_transactional | d(false)" } TASK [fedora.linux_system_roles.podman : Reboot transactional update systems] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:33 Tuesday 08 July 2025 08:54:49 -0400 (0:00:00.033) 0:00:30.932 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Fail if reboot is needed and not set] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:38 Tuesday 08 July 2025 08:54:49 -0400 (0:00:00.034) 0:00:30.967 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__podman_is_transactional | d(false)", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Get podman version] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 Tuesday 08 July 2025 08:54:49 -0400 (0:00:00.034) 0:00:31.001 ********** ok: [managed-node1] => { "changed": false, "cmd": [ "podman", "--version" ], "delta": "0:00:00.026920", "end": "2025-07-08 08:54:50.282976", "rc": 0, "start": "2025-07-08 08:54:50.256056" } STDOUT: podman version 5.5.1 TASK [fedora.linux_system_roles.podman : Set podman version] ******************* task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:52 Tuesday 08 July 2025 08:54:50 -0400 (0:00:00.385) 0:00:31.387 ********** ok: [managed-node1] => { "ansible_facts": { "podman_version": "5.5.1" }, "changed": false } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.2 or later] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:56 Tuesday 08 July 2025 08:54:50 -0400 (0:00:00.037) 0:00:31.424 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_version is version(\"4.2\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:63 Tuesday 08 July 2025 08:54:50 -0400 (0:00:00.032) 0:00:31.457 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "podman_version is version(\"4.4\", \"<\")", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 4.4 or later for quadlet, secrets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:73 Tuesday 08 July 2025 08:54:50 -0400 (0:00:00.117) 0:00:31.574 ********** META: end_host conditional evaluated to False, continuing execution for managed-node1 skipping: [managed-node1] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node1" } MSG: end_host conditional evaluated to false, continuing execution for managed-node1 TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:80 Tuesday 08 July 2025 08:54:50 -0400 (0:00:00.066) 0:00:31.641 ********** skipping: [managed-node1] => { "changed": false, "false_condition": "__has_type_pod or __has_pod_file_ext or __has_pod_file_src_ext or __has_pod_template_src_ext or __has_pod_template_src_ext_j2", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.podman : Podman package version must be 5.0 or later for Pod quadlets] *** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:96 Tuesday 08 July 2025 08:54:50 -0400 (0:00:00.062) 0:00:31.704 ********** META: end_host conditional evaluated to False, continuing execution for managed-node1 skipping: [managed-node1] => { "skip_reason": "end_host conditional evaluated to False, continuing execution for managed-node1" } MSG: end_host conditional evaluated to false, continuing execution for managed-node1 TASK [fedora.linux_system_roles.podman : Check user and group information] ***** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:109 Tuesday 08 July 2025 08:54:50 -0400 (0:00:00.065) 0:00:31.769 ********** included: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml for managed-node1 TASK [fedora.linux_system_roles.podman : Get user information] ***************** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 Tuesday 08 July 2025 08:54:50 -0400 (0:00:00.066) 0:00:31.835 ********** ok: [managed-node1] => { "ansible_facts": { "getent_passwd": { "user_quadlet_basic": null } }, "changed": false } MSG: One or more supplied key could not be found in the database. TASK [fedora.linux_system_roles.podman : Fail if user does not exist] ********** task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9 Tuesday 08 July 2025 08:54:51 -0400 (0:00:00.379) 0:00:32.215 ********** fatal: [managed-node1]: FAILED! => { "changed": false } MSG: The given podman user [user_quadlet_basic] does not exist - cannot continue TASK [Dump journal] ************************************************************ task path: /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:336 Tuesday 08 July 2025 08:54:51 -0400 (0:00:00.043) 0:00:32.258 ********** fatal: [managed-node1]: FAILED! => { "changed": false, "cmd": [ "journalctl", "-ex" ], "delta": "0:00:00.052464", "end": "2025-07-08 08:54:51.567215", "failed_when_result": true, "rc": 0, "start": "2025-07-08 08:54:51.514751" } STDOUT: Jul 08 08:51:48 managed-node1 auth_test_1_kube-auth_test_1_kube[33421]: This container is intended for podman CI testing Jul 08 08:51:48 managed-node1 podman[33425]: 2025-07-08 08:51:48.448578278 -0400 EDT m=+0.023969811 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:48 managed-node1 podman[33425]: 2025-07-08 08:51:48.463578624 -0400 EDT m=+0.038970410 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:48 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3715. Jul 08 08:51:48 managed-node1 podman[33425]: 2025-07-08 08:51:48.513685618 -0400 EDT m=+0.089077175 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:48 managed-node1 podman[33425]: 2025-07-08 08:51:48.516933005 -0400 EDT m=+0.092324634 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:48 managed-node1 auth_test_1_kube-auth_test_1_kube[33460]: This container is intended for podman CI testing Jul 08 08:51:48 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:48 managed-node1 podman[33464]: 2025-07-08 08:51:48.551624605 -0400 EDT m=+0.023398724 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:48 managed-node1 podman[33464]: 2025-07-08 08:51:48.564044867 -0400 EDT m=+0.035818947 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:48 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3720. Jul 08 08:51:48 managed-node1 podman[33464]: 2025-07-08 08:51:48.610869632 -0400 EDT m=+0.082643798 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:48 managed-node1 podman[33464]: 2025-07-08 08:51:48.614112225 -0400 EDT m=+0.085886568 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:48 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:48 managed-node1 auth_test_1_kube-auth_test_1_kube[33475]: This container is intended for podman CI testing Jul 08 08:51:48 managed-node1 podman[33479]: 2025-07-08 08:51:48.647583064 -0400 EDT m=+0.023183567 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:48 managed-node1 podman[33479]: 2025-07-08 08:51:48.661874275 -0400 EDT m=+0.037474756 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:48 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3725. Jul 08 08:51:48 managed-node1 podman[33479]: 2025-07-08 08:51:48.712876218 -0400 EDT m=+0.088476803 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:48 managed-node1 podman[33479]: 2025-07-08 08:51:48.716219436 -0400 EDT m=+0.091820092 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:48 managed-node1 auth_test_1_kube-auth_test_1_kube[33490]: This container is intended for podman CI testing Jul 08 08:51:48 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:48 managed-node1 conmon[33490]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:48 managed-node1 podman[33494]: 2025-07-08 08:51:48.751043262 -0400 EDT m=+0.023360199 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:48 managed-node1 podman[33494]: 2025-07-08 08:51:48.763434686 -0400 EDT m=+0.035751587 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:48 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3730. Jul 08 08:51:48 managed-node1 podman[33494]: 2025-07-08 08:51:48.813391895 -0400 EDT m=+0.085708857 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:48 managed-node1 auth_test_1_kube-auth_test_1_kube[33506]: This container is intended for podman CI testing Jul 08 08:51:48 managed-node1 conmon[33506]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:48 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:48 managed-node1 podman[33494]: 2025-07-08 08:51:48.817719676 -0400 EDT m=+0.090036659 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:48 managed-node1 podman[33510]: 2025-07-08 08:51:48.852289124 -0400 EDT m=+0.022895212 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:48 managed-node1 podman[33510]: 2025-07-08 08:51:48.864970582 -0400 EDT m=+0.035576641 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:48 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3735. Jul 08 08:51:48 managed-node1 podman[33510]: 2025-07-08 08:51:48.914551976 -0400 EDT m=+0.085158053 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:48 managed-node1 podman[33510]: 2025-07-08 08:51:48.917640624 -0400 EDT m=+0.088246829 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:48 managed-node1 auth_test_1_kube-auth_test_1_kube[33521]: This container is intended for podman CI testing Jul 08 08:51:48 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:48 managed-node1 podman[33525]: 2025-07-08 08:51:48.951310179 -0400 EDT m=+0.022752999 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:48 managed-node1 podman[33525]: 2025-07-08 08:51:48.963689665 -0400 EDT m=+0.035132460 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:48 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3740. Jul 08 08:51:49 managed-node1 podman[33525]: 2025-07-08 08:51:49.008565374 -0400 EDT m=+0.080008215 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:49 managed-node1 podman[33525]: 2025-07-08 08:51:49.011684347 -0400 EDT m=+0.083127174 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33537]: This container is intended for podman CI testing Jul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:49 managed-node1 podman[33541]: 2025-07-08 08:51:49.05246279 -0400 EDT m=+0.030949737 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:49 managed-node1 podman[33541]: 2025-07-08 08:51:49.067551179 -0400 EDT m=+0.046038114 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3745. Jul 08 08:51:49 managed-node1 podman[33541]: 2025-07-08 08:51:49.126159103 -0400 EDT m=+0.104646383 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33598]: This container is intended for podman CI testing Jul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:49 managed-node1 conmon[33598]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:49 managed-node1 podman[33541]: 2025-07-08 08:51:49.132373168 -0400 EDT m=+0.110860199 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:49 managed-node1 podman[33614]: 2025-07-08 08:51:49.17740571 -0400 EDT m=+0.030617289 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:49 managed-node1 podman[33614]: 2025-07-08 08:51:49.192793345 -0400 EDT m=+0.046004983 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3750. Jul 08 08:51:49 managed-node1 podman[33614]: 2025-07-08 08:51:49.248278072 -0400 EDT m=+0.101490028 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33666]: This container is intended for podman CI testing Jul 08 08:51:49 managed-node1 podman[33614]: 2025-07-08 08:51:49.253137795 -0400 EDT m=+0.106349379 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:49 managed-node1 podman[33689]: 2025-07-08 08:51:49.307534712 -0400 EDT m=+0.040092353 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:49 managed-node1 podman[33689]: 2025-07-08 08:51:49.330294602 -0400 EDT m=+0.062852173 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3755. Jul 08 08:51:49 managed-node1 podman[33689]: 2025-07-08 08:51:49.417873351 -0400 EDT m=+0.150431000 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33707]: This container is intended for podman CI testing Jul 08 08:51:49 managed-node1 conmon[33707]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:49 managed-node1 podman[33689]: 2025-07-08 08:51:49.425685791 -0400 EDT m=+0.158243377 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:49 managed-node1 python3.9[33701]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:51:49 managed-node1 podman[33711]: 2025-07-08 08:51:49.48629306 -0400 EDT m=+0.042105729 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:49 managed-node1 podman[33711]: 2025-07-08 08:51:49.505202864 -0400 EDT m=+0.061015306 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3760. Jul 08 08:51:49 managed-node1 podman[33711]: 2025-07-08 08:51:49.563285914 -0400 EDT m=+0.119098412 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:49 managed-node1 podman[33711]: 2025-07-08 08:51:49.566298505 -0400 EDT m=+0.122111008 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33752]: This container is intended for podman CI testing Jul 08 08:51:49 managed-node1 conmon[33752]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:49 managed-node1 podman[33756]: 2025-07-08 08:51:49.602293721 -0400 EDT m=+0.023104641 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:49 managed-node1 podman[33756]: 2025-07-08 08:51:49.61533105 -0400 EDT m=+0.036141951 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3765. Jul 08 08:51:49 managed-node1 podman[33756]: 2025-07-08 08:51:49.660820556 -0400 EDT m=+0.081631519 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:49 managed-node1 podman[33756]: 2025-07-08 08:51:49.663846273 -0400 EDT m=+0.084657215 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33767]: This container is intended for podman CI testing Jul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:49 managed-node1 conmon[33767]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:49 managed-node1 podman[33771]: 2025-07-08 08:51:49.696897085 -0400 EDT m=+0.021669345 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:49 managed-node1 podman[33771]: 2025-07-08 08:51:49.709430925 -0400 EDT m=+0.034203158 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3770. Jul 08 08:51:49 managed-node1 rsyslogd[811]: imjournal: journal files changed, reloading... [v8.2412.0-2.el9 try https://www.rsyslog.com/e/0 ] Jul 08 08:51:49 managed-node1 podman[33771]: 2025-07-08 08:51:49.762915854 -0400 EDT m=+0.087688178 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:49 managed-node1 podman[33771]: 2025-07-08 08:51:49.767703684 -0400 EDT m=+0.092476222 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33782]: This container is intended for podman CI testing Jul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:49 managed-node1 podman[33787]: 2025-07-08 08:51:49.800414365 -0400 EDT m=+0.022705818 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:49 managed-node1 podman[33787]: 2025-07-08 08:51:49.813402246 -0400 EDT m=+0.035693631 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3775. Jul 08 08:51:49 managed-node1 podman[33787]: 2025-07-08 08:51:49.859216266 -0400 EDT m=+0.081507708 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:49 managed-node1 podman[33787]: 2025-07-08 08:51:49.862342028 -0400 EDT m=+0.084633601 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33798]: This container is intended for podman CI testing Jul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:49 managed-node1 podman[33802]: 2025-07-08 08:51:49.894773177 -0400 EDT m=+0.022994496 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:49 managed-node1 podman[33802]: 2025-07-08 08:51:49.907297159 -0400 EDT m=+0.035518448 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3780. Jul 08 08:51:49 managed-node1 podman[33802]: 2025-07-08 08:51:49.954555528 -0400 EDT m=+0.082776810 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:49 managed-node1 podman[33802]: 2025-07-08 08:51:49.95854713 -0400 EDT m=+0.086768463 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33813]: This container is intended for podman CI testing Jul 08 08:51:49 managed-node1 conmon[33813]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:49 managed-node1 podman[33817]: 2025-07-08 08:51:49.994321706 -0400 EDT m=+0.021682296 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:50 managed-node1 podman[33817]: 2025-07-08 08:51:50.007308802 -0400 EDT m=+0.034669150 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3785. Jul 08 08:51:50 managed-node1 podman[33817]: 2025-07-08 08:51:50.06483917 -0400 EDT m=+0.092199738 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:50 managed-node1 podman[33817]: 2025-07-08 08:51:50.069602607 -0400 EDT m=+0.096962923 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[33828]: This container is intended for podman CI testing Jul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:50 managed-node1 conmon[33828]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:50 managed-node1 podman[33854]: 2025-07-08 08:51:50.124935034 -0400 EDT m=+0.038208882 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:50 managed-node1 podman[33854]: 2025-07-08 08:51:50.13785611 -0400 EDT m=+0.051130005 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3790. Jul 08 08:51:50 managed-node1 podman[33854]: 2025-07-08 08:51:50.202569983 -0400 EDT m=+0.115844060 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:50 managed-node1 podman[33854]: 2025-07-08 08:51:50.20647275 -0400 EDT m=+0.119746611 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[33918]: This container is intended for podman CI testing Jul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:50 managed-node1 podman[33940]: 2025-07-08 08:51:50.255118585 -0400 EDT m=+0.033291310 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:50 managed-node1 podman[33940]: 2025-07-08 08:51:50.26846983 -0400 EDT m=+0.046642515 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3795. Jul 08 08:51:50 managed-node1 podman[33940]: 2025-07-08 08:51:50.369192726 -0400 EDT m=+0.147365515 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[33983]: This container is intended for podman CI testing Jul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:50 managed-node1 conmon[33983]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:50 managed-node1 podman[33940]: 2025-07-08 08:51:50.376002241 -0400 EDT m=+0.154174886 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:50 managed-node1 python3.9[33978]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:51:50 managed-node1 podman[33987]: 2025-07-08 08:51:50.4323887 -0400 EDT m=+0.039929508 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:50 managed-node1 podman[33987]: 2025-07-08 08:51:50.446216676 -0400 EDT m=+0.053757208 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3800. Jul 08 08:51:50 managed-node1 podman[33987]: 2025-07-08 08:51:50.504432318 -0400 EDT m=+0.111972903 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:50 managed-node1 podman[33987]: 2025-07-08 08:51:50.507589487 -0400 EDT m=+0.115130423 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[34024]: This container is intended for podman CI testing Jul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:50 managed-node1 podman[34028]: 2025-07-08 08:51:50.543336091 -0400 EDT m=+0.023133313 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:50 managed-node1 podman[34028]: 2025-07-08 08:51:50.557285225 -0400 EDT m=+0.037082354 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3805. Jul 08 08:51:50 managed-node1 podman[34028]: 2025-07-08 08:51:50.607148898 -0400 EDT m=+0.086946107 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:50 managed-node1 podman[34028]: 2025-07-08 08:51:50.610448275 -0400 EDT m=+0.090245452 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[34039]: This container is intended for podman CI testing Jul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:50 managed-node1 podman[34043]: 2025-07-08 08:51:50.64215535 -0400 EDT m=+0.022131003 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:50 managed-node1 podman[34043]: 2025-07-08 08:51:50.655139945 -0400 EDT m=+0.035115543 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3810. Jul 08 08:51:50 managed-node1 podman[34043]: 2025-07-08 08:51:50.706854723 -0400 EDT m=+0.086830396 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:50 managed-node1 podman[34043]: 2025-07-08 08:51:50.710102295 -0400 EDT m=+0.090077940 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[34054]: This container is intended for podman CI testing Jul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:50 managed-node1 podman[34058]: 2025-07-08 08:51:50.746221367 -0400 EDT m=+0.022863699 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:50 managed-node1 podman[34058]: 2025-07-08 08:51:50.758828206 -0400 EDT m=+0.035470533 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3815. Jul 08 08:51:50 managed-node1 podman[34058]: 2025-07-08 08:51:50.808790018 -0400 EDT m=+0.085432443 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[34069]: This container is intended for podman CI testing Jul 08 08:51:50 managed-node1 conmon[34069]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:50 managed-node1 podman[34058]: 2025-07-08 08:51:50.81308862 -0400 EDT m=+0.089730948 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:50 managed-node1 podman[34073]: 2025-07-08 08:51:50.846088349 -0400 EDT m=+0.022900304 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:50 managed-node1 podman[34073]: 2025-07-08 08:51:50.858864628 -0400 EDT m=+0.035676518 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3820. Jul 08 08:51:50 managed-node1 podman[34073]: 2025-07-08 08:51:50.907379988 -0400 EDT m=+0.084191923 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:50 managed-node1 podman[34073]: 2025-07-08 08:51:50.910733502 -0400 EDT m=+0.087545474 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[34084]: This container is intended for podman CI testing Jul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:50 managed-node1 podman[34088]: 2025-07-08 08:51:50.946679837 -0400 EDT m=+0.022865188 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:50 managed-node1 podman[34088]: 2025-07-08 08:51:50.959995671 -0400 EDT m=+0.036181000 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3825. Jul 08 08:51:51 managed-node1 podman[34088]: 2025-07-08 08:51:51.002881906 -0400 EDT m=+0.079067276 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:51 managed-node1 podman[34088]: 2025-07-08 08:51:51.006032573 -0400 EDT m=+0.082218026 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34099]: This container is intended for podman CI testing Jul 08 08:51:51 managed-node1 podman[34103]: 2025-07-08 08:51:51.041650332 -0400 EDT m=+0.022409183 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:51 managed-node1 podman[34103]: 2025-07-08 08:51:51.054826908 -0400 EDT m=+0.035585744 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3830. Jul 08 08:51:51 managed-node1 podman[34103]: 2025-07-08 08:51:51.10523635 -0400 EDT m=+0.085995171 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:51 managed-node1 podman[34103]: 2025-07-08 08:51:51.108479715 -0400 EDT m=+0.089238666 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34114]: This container is intended for podman CI testing Jul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:51 managed-node1 conmon[34114]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:51 managed-node1 podman[34118]: 2025-07-08 08:51:51.145983784 -0400 EDT m=+0.023131545 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:51 managed-node1 podman[34118]: 2025-07-08 08:51:51.158714763 -0400 EDT m=+0.035862464 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3835. Jul 08 08:51:51 managed-node1 podman[34118]: 2025-07-08 08:51:51.206515345 -0400 EDT m=+0.083663128 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:51 managed-node1 podman[34118]: 2025-07-08 08:51:51.209753192 -0400 EDT m=+0.086900961 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34129]: This container is intended for podman CI testing Jul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:51 managed-node1 conmon[34129]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:51 managed-node1 podman[34133]: 2025-07-08 08:51:51.246626352 -0400 EDT m=+0.023308457 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:51 managed-node1 podman[34133]: 2025-07-08 08:51:51.259550155 -0400 EDT m=+0.036232229 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3840. Jul 08 08:51:51 managed-node1 podman[34133]: 2025-07-08 08:51:51.310043813 -0400 EDT m=+0.086725970 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:51 managed-node1 podman[34133]: 2025-07-08 08:51:51.314825939 -0400 EDT m=+0.091508123 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34144]: This container is intended for podman CI testing Jul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:51 managed-node1 conmon[34144]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:51 managed-node1 podman[34148]: 2025-07-08 08:51:51.349444872 -0400 EDT m=+0.020834015 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:51 managed-node1 podman[34148]: 2025-07-08 08:51:51.362030702 -0400 EDT m=+0.033419755 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3845. Jul 08 08:51:51 managed-node1 podman[34148]: 2025-07-08 08:51:51.411294271 -0400 EDT m=+0.082683322 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:51 managed-node1 podman[34148]: 2025-07-08 08:51:51.414607833 -0400 EDT m=+0.085996969 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34159]: This container is intended for podman CI testing Jul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:51 managed-node1 podman[34163]: 2025-07-08 08:51:51.451910069 -0400 EDT m=+0.024032297 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:51 managed-node1 podman[34163]: 2025-07-08 08:51:51.465095028 -0400 EDT m=+0.037217295 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3850. Jul 08 08:51:51 managed-node1 podman[34163]: 2025-07-08 08:51:51.51842398 -0400 EDT m=+0.090546232 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:51 managed-node1 podman[34163]: 2025-07-08 08:51:51.521671054 -0400 EDT m=+0.093793300 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34174]: This container is intended for podman CI testing Jul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:51 managed-node1 podman[34178]: 2025-07-08 08:51:51.557171448 -0400 EDT m=+0.026340600 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:51 managed-node1 podman[34178]: 2025-07-08 08:51:51.570270377 -0400 EDT m=+0.039439505 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3855. Jul 08 08:51:51 managed-node1 podman[34178]: 2025-07-08 08:51:51.623425288 -0400 EDT m=+0.092594547 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:51 managed-node1 podman[34178]: 2025-07-08 08:51:51.626725482 -0400 EDT m=+0.095894645 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34190]: This container is intended for podman CI testing Jul 08 08:51:51 managed-node1 podman[34194]: 2025-07-08 08:51:51.66428524 -0400 EDT m=+0.023796852 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:51 managed-node1 podman[34194]: 2025-07-08 08:51:51.677129052 -0400 EDT m=+0.036640614 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3860. Jul 08 08:51:51 managed-node1 podman[34194]: 2025-07-08 08:51:51.728717704 -0400 EDT m=+0.088229325 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:51 managed-node1 podman[34194]: 2025-07-08 08:51:51.731914334 -0400 EDT m=+0.091425975 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34205]: This container is intended for podman CI testing Jul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:51 managed-node1 conmon[34205]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:51 managed-node1 podman[34209]: 2025-07-08 08:51:51.766271597 -0400 EDT m=+0.022524310 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:51 managed-node1 podman[34209]: 2025-07-08 08:51:51.778901663 -0400 EDT m=+0.035154314 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3865. Jul 08 08:51:51 managed-node1 podman[34209]: 2025-07-08 08:51:51.827854907 -0400 EDT m=+0.084107688 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34220]: This container is intended for podman CI testing Jul 08 08:51:51 managed-node1 podman[34209]: 2025-07-08 08:51:51.831526913 -0400 EDT m=+0.087779765 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:51 managed-node1 conmon[34220]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:51 managed-node1 podman[34224]: 2025-07-08 08:51:51.868178779 -0400 EDT m=+0.023709243 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:51 managed-node1 podman[34224]: 2025-07-08 08:51:51.881047661 -0400 EDT m=+0.036578030 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3870. Jul 08 08:51:51 managed-node1 podman[34224]: 2025-07-08 08:51:51.93290684 -0400 EDT m=+0.088437271 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:51 managed-node1 podman[34224]: 2025-07-08 08:51:51.936102953 -0400 EDT m=+0.091633364 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34235]: This container is intended for podman CI testing Jul 08 08:51:51 managed-node1 conmon[34235]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:51 managed-node1 podman[34239]: 2025-07-08 08:51:51.974911564 -0400 EDT m=+0.023383080 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:51 managed-node1 podman[34239]: 2025-07-08 08:51:51.98759638 -0400 EDT m=+0.036067772 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3875. Jul 08 08:51:52 managed-node1 podman[34239]: 2025-07-08 08:51:52.037897828 -0400 EDT m=+0.086369361 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:52 managed-node1 podman[34239]: 2025-07-08 08:51:52.04121166 -0400 EDT m=+0.089683114 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34250]: This container is intended for podman CI testing Jul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:52 managed-node1 podman[34254]: 2025-07-08 08:51:52.076459043 -0400 EDT m=+0.022091035 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:52 managed-node1 podman[34254]: 2025-07-08 08:51:52.089098566 -0400 EDT m=+0.034730678 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3880. Jul 08 08:51:52 managed-node1 podman[34254]: 2025-07-08 08:51:52.136605391 -0400 EDT m=+0.082237394 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:52 managed-node1 podman[34254]: 2025-07-08 08:51:52.139726463 -0400 EDT m=+0.085358496 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34266]: This container is intended for podman CI testing Jul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:52 managed-node1 conmon[34266]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:52 managed-node1 podman[34270]: 2025-07-08 08:51:52.177785317 -0400 EDT m=+0.024693442 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:52 managed-node1 podman[34270]: 2025-07-08 08:51:52.190676452 -0400 EDT m=+0.037584544 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3885. Jul 08 08:51:52 managed-node1 podman[34270]: 2025-07-08 08:51:52.242920332 -0400 EDT m=+0.089828489 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34281]: This container is intended for podman CI testing Jul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:52 managed-node1 podman[34270]: 2025-07-08 08:51:52.247101011 -0400 EDT m=+0.094009235 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:52 managed-node1 conmon[34281]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:52 managed-node1 podman[34285]: 2025-07-08 08:51:52.281399672 -0400 EDT m=+0.024927817 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:52 managed-node1 podman[34285]: 2025-07-08 08:51:52.299134966 -0400 EDT m=+0.042663101 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3890. Jul 08 08:51:52 managed-node1 podman[34285]: 2025-07-08 08:51:52.343500246 -0400 EDT m=+0.087028594 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:52 managed-node1 podman[34285]: 2025-07-08 08:51:52.349245849 -0400 EDT m=+0.092773962 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34296]: This container is intended for podman CI testing Jul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:52 managed-node1 conmon[34296]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:52 managed-node1 podman[34300]: 2025-07-08 08:51:52.387822334 -0400 EDT m=+0.024037871 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:52 managed-node1 podman[34300]: 2025-07-08 08:51:52.400754882 -0400 EDT m=+0.036970368 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3895. Jul 08 08:51:52 managed-node1 podman[34300]: 2025-07-08 08:51:52.45095469 -0400 EDT m=+0.087170281 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:52 managed-node1 podman[34300]: 2025-07-08 08:51:52.454642211 -0400 EDT m=+0.090857887 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34312]: This container is intended for podman CI testing Jul 08 08:51:52 managed-node1 conmon[34312]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:52 managed-node1 podman[34316]: 2025-07-08 08:51:52.489368338 -0400 EDT m=+0.023100495 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:52 managed-node1 podman[34316]: 2025-07-08 08:51:52.501783549 -0400 EDT m=+0.035515669 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3900. Jul 08 08:51:52 managed-node1 podman[34316]: 2025-07-08 08:51:52.550546694 -0400 EDT m=+0.084278835 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:52 managed-node1 podman[34316]: 2025-07-08 08:51:52.553615063 -0400 EDT m=+0.087347258 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34327]: This container is intended for podman CI testing Jul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:52 managed-node1 conmon[34327]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:52 managed-node1 podman[34331]: 2025-07-08 08:51:52.595944229 -0400 EDT m=+0.032198277 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:52 managed-node1 podman[34331]: 2025-07-08 08:51:52.611282013 -0400 EDT m=+0.047535952 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3905. Jul 08 08:51:52 managed-node1 podman[34331]: 2025-07-08 08:51:52.669145562 -0400 EDT m=+0.105399639 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34392]: This container is intended for podman CI testing Jul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:52 managed-node1 podman[34331]: 2025-07-08 08:51:52.676901314 -0400 EDT m=+0.113155204 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:52 managed-node1 podman[34411]: 2025-07-08 08:51:52.717264706 -0400 EDT m=+0.027508917 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:52 managed-node1 podman[34411]: 2025-07-08 08:51:52.730524627 -0400 EDT m=+0.040768929 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3910. Jul 08 08:51:52 managed-node1 podman[34411]: 2025-07-08 08:51:52.786375883 -0400 EDT m=+0.096620068 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:52 managed-node1 podman[34411]: 2025-07-08 08:51:52.790202655 -0400 EDT m=+0.100446862 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34456]: This container is intended for podman CI testing Jul 08 08:51:52 managed-node1 podman[34486]: 2025-07-08 08:51:52.84544824 -0400 EDT m=+0.036617469 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:52 managed-node1 podman[34486]: 2025-07-08 08:51:52.859497057 -0400 EDT m=+0.050666352 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:52 managed-node1 python3.9[34485]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3915. Jul 08 08:51:52 managed-node1 podman[34486]: 2025-07-08 08:51:52.968714976 -0400 EDT m=+0.159884342 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34498]: This container is intended for podman CI testing Jul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:52 managed-node1 podman[34486]: 2025-07-08 08:51:52.97456631 -0400 EDT m=+0.165735868 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:53 managed-node1 podman[34504]: 2025-07-08 08:51:53.023986337 -0400 EDT m=+0.037748085 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:53 managed-node1 podman[34504]: 2025-07-08 08:51:53.037392078 -0400 EDT m=+0.051153917 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3920. Jul 08 08:51:53 managed-node1 podman[34504]: 2025-07-08 08:51:53.086015594 -0400 EDT m=+0.099777537 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:53 managed-node1 podman[34504]: 2025-07-08 08:51:53.090468412 -0400 EDT m=+0.104230229 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34539]: This container is intended for podman CI testing Jul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:53 managed-node1 conmon[34539]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:53 managed-node1 podman[34543]: 2025-07-08 08:51:53.124650233 -0400 EDT m=+0.021518634 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:53 managed-node1 podman[34543]: 2025-07-08 08:51:53.137900177 -0400 EDT m=+0.034768685 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3925. Jul 08 08:51:53 managed-node1 podman[34543]: 2025-07-08 08:51:53.189829619 -0400 EDT m=+0.086698081 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:53 managed-node1 podman[34543]: 2025-07-08 08:51:53.193012125 -0400 EDT m=+0.089880597 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34555]: This container is intended for podman CI testing Jul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:53 managed-node1 podman[34559]: 2025-07-08 08:51:53.22851641 -0400 EDT m=+0.026156436 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:53 managed-node1 podman[34559]: 2025-07-08 08:51:53.241177765 -0400 EDT m=+0.038817802 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3930. Jul 08 08:51:53 managed-node1 podman[34559]: 2025-07-08 08:51:53.28736722 -0400 EDT m=+0.085007417 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:53 managed-node1 podman[34559]: 2025-07-08 08:51:53.292001399 -0400 EDT m=+0.089641461 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34570]: This container is intended for podman CI testing Jul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:53 managed-node1 podman[34574]: 2025-07-08 08:51:53.326616039 -0400 EDT m=+0.022763830 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:53 managed-node1 podman[34574]: 2025-07-08 08:51:53.339468171 -0400 EDT m=+0.035615893 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3935. Jul 08 08:51:53 managed-node1 podman[34574]: 2025-07-08 08:51:53.381558322 -0400 EDT m=+0.077706147 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:53 managed-node1 podman[34574]: 2025-07-08 08:51:53.386169788 -0400 EDT m=+0.082317612 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34585]: This container is intended for podman CI testing Jul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:53 managed-node1 podman[34589]: 2025-07-08 08:51:53.418087203 -0400 EDT m=+0.022395652 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:53 managed-node1 podman[34589]: 2025-07-08 08:51:53.431253604 -0400 EDT m=+0.035562005 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3940. Jul 08 08:51:53 managed-node1 podman[34589]: 2025-07-08 08:51:53.477797896 -0400 EDT m=+0.082106459 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:53 managed-node1 podman[34589]: 2025-07-08 08:51:53.480897464 -0400 EDT m=+0.085205877 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34600]: This container is intended for podman CI testing Jul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:53 managed-node1 podman[34604]: 2025-07-08 08:51:53.517451555 -0400 EDT m=+0.023458453 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:53 managed-node1 podman[34604]: 2025-07-08 08:51:53.530275652 -0400 EDT m=+0.036282517 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3945. Jul 08 08:51:53 managed-node1 podman[34604]: 2025-07-08 08:51:53.575783768 -0400 EDT m=+0.081790707 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:53 managed-node1 podman[34604]: 2025-07-08 08:51:53.580302043 -0400 EDT m=+0.086308990 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34615]: This container is intended for podman CI testing Jul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:53 managed-node1 conmon[34615]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:53 managed-node1 podman[34619]: 2025-07-08 08:51:53.616636269 -0400 EDT m=+0.023020507 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:53 managed-node1 podman[34619]: 2025-07-08 08:51:53.629586441 -0400 EDT m=+0.035970587 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3950. Jul 08 08:51:53 managed-node1 podman[34619]: 2025-07-08 08:51:53.678750449 -0400 EDT m=+0.085134654 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:53 managed-node1 podman[34619]: 2025-07-08 08:51:53.681880867 -0400 EDT m=+0.088265167 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34630]: This container is intended for podman CI testing Jul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:53 managed-node1 conmon[34630]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:53 managed-node1 podman[34634]: 2025-07-08 08:51:53.717940888 -0400 EDT m=+0.022864499 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:53 managed-node1 podman[34634]: 2025-07-08 08:51:53.730421158 -0400 EDT m=+0.035344744 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3955. Jul 08 08:51:53 managed-node1 podman[34634]: 2025-07-08 08:51:53.779644519 -0400 EDT m=+0.084568160 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34645]: This container is intended for podman CI testing Jul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:53 managed-node1 podman[34634]: 2025-07-08 08:51:53.785388726 -0400 EDT m=+0.090312439 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:53 managed-node1 conmon[34645]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:53 managed-node1 podman[34649]: 2025-07-08 08:51:53.818900307 -0400 EDT m=+0.024214971 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:53 managed-node1 podman[34649]: 2025-07-08 08:51:53.83171962 -0400 EDT m=+0.037034251 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3960. Jul 08 08:51:53 managed-node1 podman[34649]: 2025-07-08 08:51:53.884257137 -0400 EDT m=+0.089571850 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:53 managed-node1 podman[34649]: 2025-07-08 08:51:53.888055587 -0400 EDT m=+0.093370276 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34661]: This container is intended for podman CI testing Jul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:53 managed-node1 podman[34665]: 2025-07-08 08:51:53.920914098 -0400 EDT m=+0.021177869 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:53 managed-node1 podman[34665]: 2025-07-08 08:51:53.935462662 -0400 EDT m=+0.035726391 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3965. Jul 08 08:51:53 managed-node1 podman[34665]: 2025-07-08 08:51:53.984659847 -0400 EDT m=+0.084923716 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:53 managed-node1 podman[34665]: 2025-07-08 08:51:53.987748479 -0400 EDT m=+0.088012382 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34676]: This container is intended for podman CI testing Jul 08 08:51:54 managed-node1 podman[34680]: 2025-07-08 08:51:54.023217593 -0400 EDT m=+0.022450999 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:54 managed-node1 podman[34680]: 2025-07-08 08:51:54.035905475 -0400 EDT m=+0.035138812 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3970. Jul 08 08:51:54 managed-node1 podman[34680]: 2025-07-08 08:51:54.085861249 -0400 EDT m=+0.085094690 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:54 managed-node1 podman[34680]: 2025-07-08 08:51:54.08899507 -0400 EDT m=+0.088228367 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34692]: This container is intended for podman CI testing Jul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:54 managed-node1 conmon[34692]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:54 managed-node1 podman[34696]: 2025-07-08 08:51:54.124730926 -0400 EDT m=+0.025816549 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:54 managed-node1 podman[34696]: 2025-07-08 08:51:54.137552729 -0400 EDT m=+0.038638342 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3975. Jul 08 08:51:54 managed-node1 podman[34696]: 2025-07-08 08:51:54.183060683 -0400 EDT m=+0.084146331 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34707]: This container is intended for podman CI testing Jul 08 08:51:54 managed-node1 podman[34696]: 2025-07-08 08:51:54.186592319 -0400 EDT m=+0.087677967 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:54 managed-node1 conmon[34707]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:54 managed-node1 podman[34711]: 2025-07-08 08:51:54.222780153 -0400 EDT m=+0.023561311 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:54 managed-node1 podman[34711]: 2025-07-08 08:51:54.235623964 -0400 EDT m=+0.036405063 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3980. Jul 08 08:51:54 managed-node1 podman[34711]: 2025-07-08 08:51:54.284151982 -0400 EDT m=+0.084933170 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34723]: This container is intended for podman CI testing Jul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:54 managed-node1 conmon[34723]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:54 managed-node1 podman[34711]: 2025-07-08 08:51:54.290397046 -0400 EDT m=+0.091178256 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:54 managed-node1 podman[34727]: 2025-07-08 08:51:54.321959858 -0400 EDT m=+0.023134363 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:54 managed-node1 podman[34727]: 2025-07-08 08:51:54.334452722 -0400 EDT m=+0.035627222 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3985. Jul 08 08:51:54 managed-node1 podman[34727]: 2025-07-08 08:51:54.382188683 -0400 EDT m=+0.083363410 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34738]: This container is intended for podman CI testing Jul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:54 managed-node1 conmon[34738]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:54 managed-node1 podman[34727]: 2025-07-08 08:51:54.389123428 -0400 EDT m=+0.090297921 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:54 managed-node1 podman[34742]: 2025-07-08 08:51:54.421808862 -0400 EDT m=+0.022864231 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:54 managed-node1 podman[34742]: 2025-07-08 08:51:54.434499651 -0400 EDT m=+0.035554851 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3990. Jul 08 08:51:54 managed-node1 podman[34742]: 2025-07-08 08:51:54.482427396 -0400 EDT m=+0.083482615 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:54 managed-node1 podman[34742]: 2025-07-08 08:51:54.485442521 -0400 EDT m=+0.086497760 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34753]: This container is intended for podman CI testing Jul 08 08:51:54 managed-node1 podman[34757]: 2025-07-08 08:51:54.519902654 -0400 EDT m=+0.021765549 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:54 managed-node1 podman[34757]: 2025-07-08 08:51:54.532609106 -0400 EDT m=+0.034471955 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 3995. Jul 08 08:51:54 managed-node1 podman[34757]: 2025-07-08 08:51:54.584264438 -0400 EDT m=+0.086127398 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:54 managed-node1 podman[34757]: 2025-07-08 08:51:54.587333117 -0400 EDT m=+0.089196094 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34768]: This container is intended for podman CI testing Jul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:54 managed-node1 conmon[34768]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:54 managed-node1 podman[34772]: 2025-07-08 08:51:54.622454787 -0400 EDT m=+0.022629409 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:54 managed-node1 podman[34772]: 2025-07-08 08:51:54.635082693 -0400 EDT m=+0.035257419 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4000. Jul 08 08:51:54 managed-node1 podman[34772]: 2025-07-08 08:51:54.68244541 -0400 EDT m=+0.082620046 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:54 managed-node1 podman[34772]: 2025-07-08 08:51:54.685503704 -0400 EDT m=+0.085678387 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34784]: This container is intended for podman CI testing Jul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:54 managed-node1 conmon[34784]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:54 managed-node1 podman[34788]: 2025-07-08 08:51:54.720325287 -0400 EDT m=+0.023012889 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:54 managed-node1 podman[34788]: 2025-07-08 08:51:54.733130807 -0400 EDT m=+0.035818363 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4005. Jul 08 08:51:54 managed-node1 podman[34788]: 2025-07-08 08:51:54.781809086 -0400 EDT m=+0.084497114 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:54 managed-node1 podman[34788]: 2025-07-08 08:51:54.784961617 -0400 EDT m=+0.087649271 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34799]: This container is intended for podman CI testing Jul 08 08:51:54 managed-node1 conmon[34799]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:54 managed-node1 podman[34803]: 2025-07-08 08:51:54.823925637 -0400 EDT m=+0.025199186 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:54 managed-node1 podman[34803]: 2025-07-08 08:51:54.837754264 -0400 EDT m=+0.039027530 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4010. Jul 08 08:51:54 managed-node1 podman[34803]: 2025-07-08 08:51:54.898369568 -0400 EDT m=+0.099642684 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34846]: This container is intended for podman CI testing Jul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:54 managed-node1 podman[34803]: 2025-07-08 08:51:54.90352077 -0400 EDT m=+0.104793920 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:54 managed-node1 conmon[34846]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:54 managed-node1 podman[34870]: 2025-07-08 08:51:54.956835937 -0400 EDT m=+0.043874313 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:54 managed-node1 podman[34870]: 2025-07-08 08:51:54.970262097 -0400 EDT m=+0.057300532 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4015. Jul 08 08:51:55 managed-node1 podman[34870]: 2025-07-08 08:51:55.030784056 -0400 EDT m=+0.117822460 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:55 managed-node1 podman[34870]: 2025-07-08 08:51:55.034551669 -0400 EDT m=+0.121590030 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:55 managed-node1 conmon[34928]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[34928]: This container is intended for podman CI testing Jul 08 08:51:55 managed-node1 podman[34949]: 2025-07-08 08:51:55.083570411 -0400 EDT m=+0.033836426 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:55 managed-node1 podman[34949]: 2025-07-08 08:51:55.101627249 -0400 EDT m=+0.051893114 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4020. Jul 08 08:51:55 managed-node1 podman[34949]: 2025-07-08 08:51:55.162314773 -0400 EDT m=+0.112580805 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[34969]: This container is intended for podman CI testing Jul 08 08:51:55 managed-node1 conmon[34969]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:55 managed-node1 podman[34949]: 2025-07-08 08:51:55.169334312 -0400 EDT m=+0.119600323 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:55 managed-node1 podman[34973]: 2025-07-08 08:51:55.2233046 -0400 EDT m=+0.039432681 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:55 managed-node1 python3.9[34962]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:51:55 managed-node1 podman[34973]: 2025-07-08 08:51:55.241773549 -0400 EDT m=+0.057901492 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4025. Jul 08 08:51:55 managed-node1 podman[34973]: 2025-07-08 08:51:55.298195371 -0400 EDT m=+0.114323417 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:55 managed-node1 podman[34973]: 2025-07-08 08:51:55.302460802 -0400 EDT m=+0.118588843 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[34999]: This container is intended for podman CI testing Jul 08 08:51:55 managed-node1 conmon[34999]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:55 managed-node1 podman[35015]: 2025-07-08 08:51:55.33487409 -0400 EDT m=+0.022966278 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:55 managed-node1 podman[35015]: 2025-07-08 08:51:55.347928004 -0400 EDT m=+0.036020177 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4030. Jul 08 08:51:55 managed-node1 podman[35015]: 2025-07-08 08:51:55.39641325 -0400 EDT m=+0.084505554 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:55 managed-node1 podman[35015]: 2025-07-08 08:51:55.399632598 -0400 EDT m=+0.087724869 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35026]: This container is intended for podman CI testing Jul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:55 managed-node1 conmon[35026]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:55 managed-node1 podman[35030]: 2025-07-08 08:51:55.432369825 -0400 EDT m=+0.021888850 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:55 managed-node1 podman[35030]: 2025-07-08 08:51:55.444847265 -0400 EDT m=+0.034366252 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4035. Jul 08 08:51:55 managed-node1 podman[35030]: 2025-07-08 08:51:55.494822879 -0400 EDT m=+0.084342000 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:55 managed-node1 podman[35030]: 2025-07-08 08:51:55.49802887 -0400 EDT m=+0.087547889 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35041]: This container is intended for podman CI testing Jul 08 08:51:55 managed-node1 conmon[35041]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:55 managed-node1 podman[35045]: 2025-07-08 08:51:55.534978267 -0400 EDT m=+0.022300939 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:55 managed-node1 podman[35045]: 2025-07-08 08:51:55.547819511 -0400 EDT m=+0.035142129 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4040. Jul 08 08:51:55 managed-node1 podman[35045]: 2025-07-08 08:51:55.596409489 -0400 EDT m=+0.083732130 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:55 managed-node1 podman[35045]: 2025-07-08 08:51:55.599513007 -0400 EDT m=+0.086835856 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35056]: This container is intended for podman CI testing Jul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:55 managed-node1 podman[35060]: 2025-07-08 08:51:55.635383437 -0400 EDT m=+0.022862791 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:55 managed-node1 podman[35060]: 2025-07-08 08:51:55.647910671 -0400 EDT m=+0.035389965 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4045. Jul 08 08:51:55 managed-node1 podman[35060]: 2025-07-08 08:51:55.691636435 -0400 EDT m=+0.079115737 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:55 managed-node1 podman[35060]: 2025-07-08 08:51:55.695110225 -0400 EDT m=+0.082589576 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:55 managed-node1 conmon[35071]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35071]: This container is intended for podman CI testing Jul 08 08:51:55 managed-node1 podman[35075]: 2025-07-08 08:51:55.728779782 -0400 EDT m=+0.022198450 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:55 managed-node1 podman[35075]: 2025-07-08 08:51:55.741390318 -0400 EDT m=+0.034808950 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4050. Jul 08 08:51:55 managed-node1 podman[35075]: 2025-07-08 08:51:55.789562384 -0400 EDT m=+0.082981033 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:55 managed-node1 podman[35075]: 2025-07-08 08:51:55.793529932 -0400 EDT m=+0.086948620 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35086]: This container is intended for podman CI testing Jul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:55 managed-node1 podman[35090]: 2025-07-08 08:51:55.830769937 -0400 EDT m=+0.023577778 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:55 managed-node1 podman[35090]: 2025-07-08 08:51:55.843324486 -0400 EDT m=+0.036132283 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4055. Jul 08 08:51:55 managed-node1 podman[35090]: 2025-07-08 08:51:55.894903684 -0400 EDT m=+0.087711555 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:55 managed-node1 podman[35090]: 2025-07-08 08:51:55.898507737 -0400 EDT m=+0.091315579 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35101]: This container is intended for podman CI testing Jul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:55 managed-node1 conmon[35101]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:55 managed-node1 podman[35105]: 2025-07-08 08:51:55.930536412 -0400 EDT m=+0.021767162 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:55 managed-node1 podman[35105]: 2025-07-08 08:51:55.943827468 -0400 EDT m=+0.035058234 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4060. Jul 08 08:51:55 managed-node1 podman[35105]: 2025-07-08 08:51:55.992851538 -0400 EDT m=+0.084082447 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:55 managed-node1 podman[35105]: 2025-07-08 08:51:55.996511805 -0400 EDT m=+0.087742643 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35117]: This container is intended for podman CI testing Jul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:55 managed-node1 conmon[35117]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:56 managed-node1 podman[35121]: 2025-07-08 08:51:56.032338614 -0400 EDT m=+0.022974943 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:56 managed-node1 podman[35121]: 2025-07-08 08:51:56.044592593 -0400 EDT m=+0.035228838 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4065. Jul 08 08:51:56 managed-node1 podman[35121]: 2025-07-08 08:51:56.092083577 -0400 EDT m=+0.082719908 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35133]: This container is intended for podman CI testing Jul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:56 managed-node1 conmon[35133]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:56 managed-node1 podman[35121]: 2025-07-08 08:51:56.100143213 -0400 EDT m=+0.090779556 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:56 managed-node1 podman[35137]: 2025-07-08 08:51:56.1295111 -0400 EDT m=+0.023708974 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:56 managed-node1 podman[35137]: 2025-07-08 08:51:56.142683417 -0400 EDT m=+0.036881232 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4070. Jul 08 08:51:56 managed-node1 podman[35137]: 2025-07-08 08:51:56.190097155 -0400 EDT m=+0.084295035 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35148]: This container is intended for podman CI testing Jul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:56 managed-node1 podman[35137]: 2025-07-08 08:51:56.194245804 -0400 EDT m=+0.088443924 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:56 managed-node1 podman[35152]: 2025-07-08 08:51:56.226636924 -0400 EDT m=+0.023968011 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:56 managed-node1 podman[35152]: 2025-07-08 08:51:56.239370854 -0400 EDT m=+0.036701891 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4075. Jul 08 08:51:56 managed-node1 podman[35152]: 2025-07-08 08:51:56.28775851 -0400 EDT m=+0.085089601 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:56 managed-node1 podman[35152]: 2025-07-08 08:51:56.290845272 -0400 EDT m=+0.088176435 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35163]: This container is intended for podman CI testing Jul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:56 managed-node1 conmon[35163]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:56 managed-node1 podman[35167]: 2025-07-08 08:51:56.326445838 -0400 EDT m=+0.023013479 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:56 managed-node1 podman[35167]: 2025-07-08 08:51:56.33903617 -0400 EDT m=+0.035603761 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4080. Jul 08 08:51:56 managed-node1 podman[35167]: 2025-07-08 08:51:56.392289413 -0400 EDT m=+0.088857050 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35178]: This container is intended for podman CI testing Jul 08 08:51:56 managed-node1 podman[35167]: 2025-07-08 08:51:56.396104065 -0400 EDT m=+0.092671932 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:56 managed-node1 conmon[35178]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:56 managed-node1 podman[35182]: 2025-07-08 08:51:56.430712998 -0400 EDT m=+0.021670054 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:56 managed-node1 podman[35182]: 2025-07-08 08:51:56.443484263 -0400 EDT m=+0.034441280 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4085. Jul 08 08:51:56 managed-node1 podman[35182]: 2025-07-08 08:51:56.492719275 -0400 EDT m=+0.083676318 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:56 managed-node1 podman[35182]: 2025-07-08 08:51:56.495965431 -0400 EDT m=+0.086922524 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35193]: This container is intended for podman CI testing Jul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:56 managed-node1 conmon[35193]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:56 managed-node1 podman[35197]: 2025-07-08 08:51:56.536373017 -0400 EDT m=+0.027245791 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:56 managed-node1 podman[35197]: 2025-07-08 08:51:56.549889706 -0400 EDT m=+0.040762533 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4090. Jul 08 08:51:56 managed-node1 podman[35197]: 2025-07-08 08:51:56.610056147 -0400 EDT m=+0.100928950 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35231]: This container is intended for podman CI testing Jul 08 08:51:56 managed-node1 conmon[35231]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:56 managed-node1 podman[35197]: 2025-07-08 08:51:56.617514279 -0400 EDT m=+0.108386978 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:56 managed-node1 podman[35264]: 2025-07-08 08:51:56.669979715 -0400 EDT m=+0.043157775 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:56 managed-node1 podman[35264]: 2025-07-08 08:51:56.68324339 -0400 EDT m=+0.056421529 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4095. Jul 08 08:51:56 managed-node1 podman[35264]: 2025-07-08 08:51:56.741763648 -0400 EDT m=+0.114941702 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:56 managed-node1 podman[35264]: 2025-07-08 08:51:56.745717027 -0400 EDT m=+0.118895041 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35322]: This container is intended for podman CI testing Jul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:56 managed-node1 conmon[35322]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:56 managed-node1 podman[35351]: 2025-07-08 08:51:56.804720401 -0400 EDT m=+0.043304467 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:56 managed-node1 podman[35351]: 2025-07-08 08:51:56.818086586 -0400 EDT m=+0.056670229 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:56 managed-node1 python3.9[35352]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4100. Jul 08 08:51:56 managed-node1 podman[35351]: 2025-07-08 08:51:56.921640357 -0400 EDT m=+0.160224188 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35364]: This container is intended for podman CI testing Jul 08 08:51:56 managed-node1 conmon[35364]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:56 managed-node1 podman[35351]: 2025-07-08 08:51:56.929562868 -0400 EDT m=+0.168146477 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:56 managed-node1 podman[35370]: 2025-07-08 08:51:56.978581974 -0400 EDT m=+0.040551987 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:56 managed-node1 podman[35370]: 2025-07-08 08:51:56.991489009 -0400 EDT m=+0.053458984 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4105. Jul 08 08:51:57 managed-node1 podman[35370]: 2025-07-08 08:51:57.041682588 -0400 EDT m=+0.103652603 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35405]: This container is intended for podman CI testing Jul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:57 managed-node1 podman[35370]: 2025-07-08 08:51:57.046108774 -0400 EDT m=+0.108078815 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:57 managed-node1 podman[35409]: 2025-07-08 08:51:57.080657767 -0400 EDT m=+0.025552325 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:57 managed-node1 podman[35409]: 2025-07-08 08:51:57.093888586 -0400 EDT m=+0.038783111 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4110. Jul 08 08:51:57 managed-node1 podman[35409]: 2025-07-08 08:51:57.139013153 -0400 EDT m=+0.083907706 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35420]: This container is intended for podman CI testing Jul 08 08:51:57 managed-node1 podman[35409]: 2025-07-08 08:51:57.144376284 -0400 EDT m=+0.089270897 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:57 managed-node1 podman[35424]: 2025-07-08 08:51:57.181570699 -0400 EDT m=+0.027520626 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:57 managed-node1 podman[35424]: 2025-07-08 08:51:57.194248987 -0400 EDT m=+0.040198878 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4115. Jul 08 08:51:57 managed-node1 podman[35424]: 2025-07-08 08:51:57.242720898 -0400 EDT m=+0.088670798 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:57 managed-node1 podman[35424]: 2025-07-08 08:51:57.245785923 -0400 EDT m=+0.091735824 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35435]: This container is intended for podman CI testing Jul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:57 managed-node1 podman[35439]: 2025-07-08 08:51:57.282621332 -0400 EDT m=+0.023409359 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:57 managed-node1 podman[35439]: 2025-07-08 08:51:57.295172051 -0400 EDT m=+0.035960073 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4120. Jul 08 08:51:57 managed-node1 podman[35439]: 2025-07-08 08:51:57.34336529 -0400 EDT m=+0.084153346 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:57 managed-node1 podman[35439]: 2025-07-08 08:51:57.346573648 -0400 EDT m=+0.087361807 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35450]: This container is intended for podman CI testing Jul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:57 managed-node1 podman[35454]: 2025-07-08 08:51:57.379900846 -0400 EDT m=+0.022316677 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:57 managed-node1 podman[35454]: 2025-07-08 08:51:57.392561635 -0400 EDT m=+0.034977428 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4125. Jul 08 08:51:57 managed-node1 podman[35454]: 2025-07-08 08:51:57.450331168 -0400 EDT m=+0.092747058 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:57 managed-node1 podman[35454]: 2025-07-08 08:51:57.453552372 -0400 EDT m=+0.095968236 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35465]: This container is intended for podman CI testing Jul 08 08:51:57 managed-node1 conmon[35465]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:57 managed-node1 podman[35469]: 2025-07-08 08:51:57.489276437 -0400 EDT m=+0.023295529 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:57 managed-node1 podman[35469]: 2025-07-08 08:51:57.501975537 -0400 EDT m=+0.035994590 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4130. Jul 08 08:51:57 managed-node1 podman[35469]: 2025-07-08 08:51:57.550665142 -0400 EDT m=+0.084684211 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:57 managed-node1 podman[35469]: 2025-07-08 08:51:57.553825967 -0400 EDT m=+0.087845107 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35480]: This container is intended for podman CI testing Jul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:57 managed-node1 podman[35484]: 2025-07-08 08:51:57.58960223 -0400 EDT m=+0.022372847 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:57 managed-node1 podman[35484]: 2025-07-08 08:51:57.605935724 -0400 EDT m=+0.038706383 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4135. Jul 08 08:51:57 managed-node1 podman[35484]: 2025-07-08 08:51:57.665840337 -0400 EDT m=+0.098611026 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:57 managed-node1 podman[35484]: 2025-07-08 08:51:57.669206337 -0400 EDT m=+0.101977039 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35521]: This container is intended for podman CI testing Jul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:57 managed-node1 podman[35551]: 2025-07-08 08:51:57.723078909 -0400 EDT m=+0.044059134 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:57 managed-node1 podman[35551]: 2025-07-08 08:51:57.738242838 -0400 EDT m=+0.059223139 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4140. Jul 08 08:51:57 managed-node1 podman[35551]: 2025-07-08 08:51:57.795605487 -0400 EDT m=+0.116585731 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:57 managed-node1 podman[35551]: 2025-07-08 08:51:57.802164945 -0400 EDT m=+0.123145103 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35610]: This container is intended for podman CI testing Jul 08 08:51:57 managed-node1 conmon[35610]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:57 managed-node1 podman[35640]: 2025-07-08 08:51:57.861994014 -0400 EDT m=+0.040045330 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:57 managed-node1 podman[35640]: 2025-07-08 08:51:57.878110989 -0400 EDT m=+0.056162178 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4145. Jul 08 08:51:57 managed-node1 podman[35640]: 2025-07-08 08:51:57.934381076 -0400 EDT m=+0.112432215 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35652]: This container is intended for podman CI testing Jul 08 08:51:57 managed-node1 conmon[35652]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:57 managed-node1 podman[35640]: 2025-07-08 08:51:57.942413628 -0400 EDT m=+0.120464749 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:57 managed-node1 python3.9[35639]: ansible-ansible.legacy.command Invoked with _raw_params=systemd-escape --template podman-kube@.service /etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:51:57 managed-node1 podman[35656]: 2025-07-08 08:51:57.99609442 -0400 EDT m=+0.039690418 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:58 managed-node1 podman[35656]: 2025-07-08 08:51:58.015367809 -0400 EDT m=+0.058963962 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:58 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4150. Jul 08 08:51:58 managed-node1 podman[35656]: 2025-07-08 08:51:58.071702173 -0400 EDT m=+0.115298222 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:58 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:58 managed-node1 auth_test_1_kube-auth_test_1_kube[35691]: This container is intended for podman CI testing Jul 08 08:51:58 managed-node1 podman[35656]: 2025-07-08 08:51:58.081398246 -0400 EDT m=+0.124994462 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:58 managed-node1 podman[35696]: 2025-07-08 08:51:58.118603324 -0400 EDT m=+0.029031609 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:58 managed-node1 podman[35696]: 2025-07-08 08:51:58.132892828 -0400 EDT m=+0.043321060 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:58 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4155. Jul 08 08:51:58 managed-node1 podman[35696]: 2025-07-08 08:51:58.179673011 -0400 EDT m=+0.090101285 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:58 managed-node1 podman[35696]: 2025-07-08 08:51:58.182883548 -0400 EDT m=+0.093311839 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:58 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:58 managed-node1 auth_test_1_kube-auth_test_1_kube[35707]: This container is intended for podman CI testing Jul 08 08:51:58 managed-node1 conmon[35707]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:58 managed-node1 podman[35712]: 2025-07-08 08:51:58.228222867 -0400 EDT m=+0.032629915 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:58 managed-node1 podman[35712]: 2025-07-08 08:51:58.241845369 -0400 EDT m=+0.046252354 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:58 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4160. Jul 08 08:51:58 managed-node1 podman[35712]: 2025-07-08 08:51:58.306194824 -0400 EDT m=+0.110602009 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:58 managed-node1 podman[35712]: 2025-07-08 08:51:58.311409402 -0400 EDT m=+0.115816394 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:58 managed-node1 auth_test_1_kube-auth_test_1_kube[35775]: This container is intended for podman CI testing Jul 08 08:51:58 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:58 managed-node1 podman[35800]: 2025-07-08 08:51:58.350154057 -0400 EDT m=+0.026407392 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:58 managed-node1 podman[35800]: 2025-07-08 08:51:58.365884381 -0400 EDT m=+0.042137951 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:58 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4165. Jul 08 08:51:58 managed-node1 podman[35800]: 2025-07-08 08:51:58.428581949 -0400 EDT m=+0.104835493 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:58 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:58 managed-node1 auth_test_1_kube-auth_test_1_kube[35844]: This container is intended for podman CI testing Jul 08 08:51:58 managed-node1 conmon[35844]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:58 managed-node1 podman[35800]: 2025-07-08 08:51:58.436307291 -0400 EDT m=+0.112560702 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:58 managed-node1 podman[35867]: 2025-07-08 08:51:58.487344628 -0400 EDT m=+0.043335121 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:58 managed-node1 podman[35867]: 2025-07-08 08:51:58.500661958 -0400 EDT m=+0.056652344 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:58 managed-node1 python3.9[35865]: ansible-systemd Invoked with name=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service scope=system state=stopped enabled=False daemon_reload=False daemon_reexec=False no_block=False force=None masked=None Jul 08 08:51:58 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4170. Jul 08 08:51:58 managed-node1 podman[35867]: 2025-07-08 08:51:58.703250209 -0400 EDT m=+0.259240736 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:58 managed-node1 auth_test_1_kube-auth_test_1_kube[35878]: This container is intended for podman CI testing Jul 08 08:51:58 managed-node1 podman[35867]: 2025-07-08 08:51:58.708624472 -0400 EDT m=+0.264614833 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:58 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:58 managed-node1 systemd[1]: Reloading. Jul 08 08:51:58 managed-node1 podman[35884]: 2025-07-08 08:51:58.78180345 -0400 EDT m=+0.056786185 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:58 managed-node1 podman[35884]: 2025-07-08 08:51:58.796379401 -0400 EDT m=+0.071362044 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test) Jul 08 08:51:58 managed-node1 systemd-rc-local-generator[35907]: /etc/rc.d/rc.local is not marked executable, skipping. Jul 08 08:51:58 managed-node1 systemd[1]: Stopping A template for running K8s workloads via podman-kube-play... â–‘â–‘ Subject: A stop job for unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has begun execution â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A stop job for unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has begun execution. â–‘â–‘ â–‘â–‘ The job identifier is 4175. Jul 08 08:51:59 managed-node1 systemd[1]: Started libcrun container. â–‘â–‘ Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4176. Jul 08 08:51:59 managed-node1 podman[35884]: 2025-07-08 08:51:59.019304564 -0400 EDT m=+0.294287240 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0) Jul 08 08:51:59 managed-node1 podman[35884]: 2025-07-08 08:51:59.022879659 -0400 EDT m=+0.297862353 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage) Jul 08 08:51:59 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state. Jul 08 08:51:59 managed-node1 auth_test_1_kube-auth_test_1_kube[35932]: This container is intended for podman CI testing Jul 08 08:51:59 managed-node1 conmon[35932]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events Jul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.048700954 -0400 EDT m=+0.061178515 pod stop 8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c (image=, name=auth_test_1_kube) Jul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.050394331 -0400 EDT m=+0.062872020 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:59 managed-node1 systemd[1]: var-lib-containers-storage-overlay-765a43272b7d1ee027833f1b9208178861d0063df0fe51067dc546f0ffe0a36b-merged.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay-765a43272b7d1ee027833f1b9208178861d0063df0fe51067dc546f0ffe0a36b-merged.mount has successfully entered the 'dead' state. Jul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.096509549 -0400 EDT m=+0.108986982 container cleanup e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry) Jul 08 08:51:59 managed-node1 systemd[1]: libpod-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a.scope has successfully entered the 'dead' state. Jul 08 08:51:59 managed-node1 conmon[20816]: conmon 83c6924d2fc59b99e53d : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a.scope/container/memory.events Jul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.117019225 -0400 EDT m=+0.129496774 container died 83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a (image=, name=8e1f57b75507-infra, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:59 managed-node1 systemd[1]: run-r453f02fb4ae24dce84d3e8de47463c2f.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit run-r453f02fb4ae24dce84d3e8de47463c2f.scope has successfully entered the 'dead' state. Jul 08 08:51:59 managed-node1 kernel: podman1: port 1(veth1) entered disabled state Jul 08 08:51:59 managed-node1 kernel: veth1 (unregistering): left allmulticast mode Jul 08 08:51:59 managed-node1 kernel: veth1 (unregistering): left promiscuous mode Jul 08 08:51:59 managed-node1 kernel: podman1: port 1(veth1) entered disabled state Jul 08 08:51:59 managed-node1 NetworkManager[642]: [1751979119.1744] device (podman1): state change: activated -> unmanaged (reason 'unmanaged', managed-type: 'removed') Jul 08 08:51:59 managed-node1 systemd[1]: Starting Network Manager Script Dispatcher Service... â–‘â–‘ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit NetworkManager-dispatcher.service has begun execution. â–‘â–‘ â–‘â–‘ The job identifier is 4182. Jul 08 08:51:59 managed-node1 systemd[1]: Started Network Manager Script Dispatcher Service. â–‘â–‘ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit NetworkManager-dispatcher.service has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4182. Jul 08 08:51:59 managed-node1 systemd[1]: run-netns-netns\x2d776ce46f\x2dcdf2\x2d9675\x2daf23\x2dbc7393892a77.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit run-netns-netns\x2d776ce46f\x2dcdf2\x2d9675\x2daf23\x2dbc7393892a77.mount has successfully entered the 'dead' state. Jul 08 08:51:59 managed-node1 systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a-rootfs-merge.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay\x2dcontainers-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a-rootfs-merge.mount has successfully entered the 'dead' state. Jul 08 08:51:59 managed-node1 systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a-userdata-shm.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay\x2dcontainers-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a-userdata-shm.mount has successfully entered the 'dead' state. Jul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.309923758 -0400 EDT m=+0.322401287 container cleanup 83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a (image=, name=8e1f57b75507-infra, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:59 managed-node1 systemd[1]: Removed slice cgroup machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice. â–‘â–‘ Subject: A stop job for unit machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice has finished â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A stop job for unit machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice has finished. â–‘â–‘ â–‘â–‘ The job identifier is 4248 and the job result is done. Jul 08 08:51:59 managed-node1 systemd[1]: machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice: Consumed 1.763s CPU time. â–‘â–‘ Subject: Resources consumed by unit runtime â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice completed and consumed the indicated resources. Jul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.33731008 -0400 EDT m=+0.349787550 container remove e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z) Jul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.365548153 -0400 EDT m=+0.378025627 container remove 83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a (image=, name=8e1f57b75507-infra, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:59 managed-node1 systemd[1]: machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice: Failed to open /run/systemd/transient/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice: No such file or directory Jul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.37391867 -0400 EDT m=+0.386396101 pod remove 8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c (image=, name=auth_test_1_kube) Jul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.377980277 -0400 EDT m=+0.390457910 container kill 8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1 (image=, name=3ef6fcac6278-service, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:59 managed-node1 systemd[1]: libpod-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1.scope has successfully entered the 'dead' state. Jul 08 08:51:59 managed-node1 conmon[20738]: conmon 8ab849217cf3dcfe0949 : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/libpod-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1.scope/container/memory.events Jul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.384674549 -0400 EDT m=+0.397152102 container died 8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1 (image=, name=3ef6fcac6278-service, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:59 managed-node1 systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1-rootfs-merge.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay\x2dcontainers-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1-rootfs-merge.mount has successfully entered the 'dead' state. Jul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.442987705 -0400 EDT m=+0.455465175 container remove 8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1 (image=, name=3ef6fcac6278-service, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service) Jul 08 08:51:59 managed-node1 podman[35930]: Pods stopped: Jul 08 08:51:59 managed-node1 podman[35930]: 8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c Jul 08 08:51:59 managed-node1 podman[35930]: Pods removed: Jul 08 08:51:59 managed-node1 podman[35930]: 8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c Jul 08 08:51:59 managed-node1 podman[35930]: Secrets removed: Jul 08 08:51:59 managed-node1 podman[35930]: Volumes removed: Jul 08 08:51:59 managed-node1 systemd[1]: podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has successfully entered the 'dead' state. Jul 08 08:51:59 managed-node1 systemd[1]: Stopped A template for running K8s workloads via podman-kube-play. â–‘â–‘ Subject: A stop job for unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has finished â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A stop job for unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service has finished. â–‘â–‘ â–‘â–‘ The job identifier is 4175 and the job result is done. Jul 08 08:51:59 managed-node1 systemd[1]: podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service: Consumed 37.552s CPU time. â–‘â–‘ Subject: Resources consumed by unit runtime â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service completed and consumed the indicated resources. Jul 08 08:51:59 managed-node1 python3.9[36157]: ansible-stat Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:00 managed-node1 systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1-userdata-shm.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay\x2dcontainers-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1-userdata-shm.mount has successfully entered the 'dead' state. Jul 08 08:52:00 managed-node1 python3.9[36308]: ansible-containers.podman.podman_play Invoked with state=absent kube_file=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml executable=podman annotation=None kube_file_content=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None debug=None quiet=None recreate=None userns=None log_level=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None Jul 08 08:52:00 managed-node1 python3.9[36308]: ansible-containers.podman.podman_play version: 5.5.1, kube file /etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml Jul 08 08:52:00 managed-node1 python3.9[36471]: ansible-file Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:02 managed-node1 python3.9[36620]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:03 managed-node1 python3.9[36771]: ansible-systemd Invoked with name=auth_test_1_quadlet.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None Jul 08 08:52:03 managed-node1 systemd[1]: Reloading. Jul 08 08:52:03 managed-node1 systemd-rc-local-generator[36791]: /etc/rc.d/rc.local is not marked executable, skipping. Jul 08 08:52:03 managed-node1 python3.9[36956]: ansible-stat Invoked with path=/etc/containers/systemd/auth_test_1_quadlet.container follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:04 managed-node1 python3.9[37256]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:07 managed-node1 python3.9[37555]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:08 managed-node1 python3.9[37710]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:09 managed-node1 systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Jul 08 08:52:10 managed-node1 python3.9[37861]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:12 managed-node1 python3.9[38012]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:14 managed-node1 python3.9[38163]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:15 managed-node1 python3.9[38314]: ansible-ansible.legacy.command Invoked with _raw_params=systemd-escape --template podman-kube@.service /etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:16 managed-node1 python3.9[38464]: ansible-systemd Invoked with name=podman-kube@-etc-containers-ansible\x2dkubernetes.d-auth_test_1_kube.yml.service scope=system state=stopped enabled=False daemon_reload=False daemon_reexec=False no_block=False force=None masked=None Jul 08 08:52:17 managed-node1 python3.9[38615]: ansible-stat Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:17 managed-node1 python3.9[38764]: ansible-file Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:19 managed-node1 python3.9[38913]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:20 managed-node1 python3.9[39064]: ansible-systemd Invoked with name=auth_test_1_quadlet.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None Jul 08 08:52:20 managed-node1 systemd[1]: Reloading. Jul 08 08:52:20 managed-node1 systemd-rc-local-generator[39087]: /etc/rc.d/rc.local is not marked executable, skipping. Jul 08 08:52:21 managed-node1 python3.9[39249]: ansible-stat Invoked with path=/etc/containers/systemd/auth_test_1_quadlet.container follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:22 managed-node1 python3.9[39549]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:24 managed-node1 python3.9[39848]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:25 managed-node1 python3.9[40003]: ansible-getent Invoked with database=passwd key=auth_test_user1 fail_key=False service=None split=None Jul 08 08:52:25 managed-node1 python3.9[40153]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:26 managed-node1 python3.9[40303]: ansible-user Invoked with name=auth_test_user1 state=absent non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on managed-node1 update_password=always uid=None group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None password_expire_warn=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None Jul 08 08:52:26 managed-node1 python3.9[40453]: ansible-file Invoked with path=/home/auth_test_user1 state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:27 managed-node1 python3.9[40602]: ansible-ansible.legacy.command Invoked with _raw_params=podman inspect podman_registry --format '{{range .}}{{range .Mounts}}{{if eq .Type "volume"}}{{.Name}}{{end}}{{end}}{{end}}' _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:27 managed-node1 python3.9[40759]: ansible-ansible.legacy.command Invoked with _raw_params=podman rm -f podman_registry _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:27 managed-node1 systemd[1]: libpod-2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354.scope has successfully entered the 'dead' state. Jul 08 08:52:27 managed-node1 podman[40760]: 2025-07-08 08:52:27.930931198 -0400 EDT m=+0.044482479 container died 2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354 (image=quay.io/libpod/registry:2.8.2, name=podman_registry) Jul 08 08:52:27 managed-node1 kernel: podman0: port 1(veth0) entered disabled state Jul 08 08:52:27 managed-node1 kernel: veth0 (unregistering): left allmulticast mode Jul 08 08:52:27 managed-node1 kernel: veth0 (unregistering): left promiscuous mode Jul 08 08:52:27 managed-node1 kernel: podman0: port 1(veth0) entered disabled state Jul 08 08:52:27 managed-node1 NetworkManager[642]: [1751979147.9779] device (podman0): state change: activated -> unmanaged (reason 'unmanaged', managed-type: 'removed') Jul 08 08:52:27 managed-node1 systemd[1]: Starting Network Manager Script Dispatcher Service... â–‘â–‘ Subject: A start job for unit NetworkManager-dispatcher.service has begun execution â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit NetworkManager-dispatcher.service has begun execution. â–‘â–‘ â–‘â–‘ The job identifier is 4250. Jul 08 08:52:27 managed-node1 systemd[1]: Started Network Manager Script Dispatcher Service. â–‘â–‘ Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ A start job for unit NetworkManager-dispatcher.service has finished successfully. â–‘â–‘ â–‘â–‘ The job identifier is 4250. Jul 08 08:52:28 managed-node1 systemd[1]: run-netns-netns\x2dcbdb4e71\x2d6d48\x2da1b0\x2d498d\x2d53f813dda638.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit run-netns-netns\x2dcbdb4e71\x2d6d48\x2da1b0\x2d498d\x2d53f813dda638.mount has successfully entered the 'dead' state. Jul 08 08:52:28 managed-node1 systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354-userdata-shm.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay\x2dcontainers-2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354-userdata-shm.mount has successfully entered the 'dead' state. Jul 08 08:52:28 managed-node1 systemd[1]: var-lib-containers-storage-overlay-7ad31fb747dc13efb3b22f585d1a06b75296c677706679c325a915f2d85e1e11-merged.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay-7ad31fb747dc13efb3b22f585d1a06b75296c677706679c325a915f2d85e1e11-merged.mount has successfully entered the 'dead' state. Jul 08 08:52:28 managed-node1 podman[40760]: 2025-07-08 08:52:28.143410392 -0400 EDT m=+0.256961352 container remove 2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354 (image=quay.io/libpod/registry:2.8.2, name=podman_registry) Jul 08 08:52:28 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 08 08:52:28 managed-node1 systemd[1]: libpod-conmon-2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354.scope: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit libpod-conmon-2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354.scope has successfully entered the 'dead' state. Jul 08 08:52:28 managed-node1 python3.9[40973]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume rm c5dabb915f4c30f58eb29c345d18cc09b42d61980a614c3a9fecfd122d7d7dbb _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:28 managed-node1 podman[40974]: 2025-07-08 08:52:28.605123907 -0400 EDT m=+0.025267136 volume remove c5dabb915f4c30f58eb29c345d18cc09b42d61980a614c3a9fecfd122d7d7dbb Jul 08 08:52:28 managed-node1 python3.9[41130]: ansible-file Invoked with path=/tmp/lsr_dbjlhh8o_podman state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:29 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 08 08:52:31 managed-node1 python3.9[41328]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Jul 08 08:52:32 managed-node1 python3.9[41503]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:32 managed-node1 python3.9[41652]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:34 managed-node1 python3.9[41950]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:35 managed-node1 python3.9[42105]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Jul 08 08:52:36 managed-node1 python3.9[42255]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:38 managed-node1 systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Jul 08 08:52:38 managed-node1 python3.9[42406]: ansible-tempfile Invoked with state=directory prefix=lsr_podman_config_ suffix= path=None Jul 08 08:52:38 managed-node1 python3.9[42555]: ansible-ansible.legacy.command Invoked with _raw_params=tar --ignore-failed-read -c -P -v -p -f /tmp/lsr_podman_config_6ihb6fh6/backup.tar /etc/containers/containers.conf.d/50-systemroles.conf /etc/containers/registries.conf.d/50-systemroles.conf /etc/containers/storage.conf /etc/containers/policy.json _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:39 managed-node1 python3.9[42705]: ansible-user Invoked with name=user1 state=present non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on managed-node1 update_password=always uid=None group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None password_expire_warn=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None Jul 08 08:52:39 managed-node1 useradd[42707]: new group: name=user1, GID=1000 Jul 08 08:52:39 managed-node1 useradd[42707]: new user: name=user1, UID=1000, GID=1000, home=/home/user1, shell=/bin/bash, from=/dev/pts/0 Jul 08 08:52:41 managed-node1 python3.9[43011]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:42 managed-node1 python3.9[43167]: ansible-getent Invoked with database=passwd key=user1 fail_key=False service=None split=None Jul 08 08:52:42 managed-node1 python3.9[43317]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:42 managed-node1 python3.9[43468]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:43 managed-node1 python3.9[43618]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:44 managed-node1 python3.9[43768]: ansible-file Invoked with path=/home/user1/.config/containers/containers.conf.d state=directory owner=user1 group=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:44 managed-node1 python3.9[43917]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:52:45 managed-node1 python3.9[44037]: ansible-ansible.legacy.copy Invoked with dest=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf owner=user1 group=user1 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979164.4055185-13192-99143701796826/.source.conf _original_basename=.fz16347w follow=False checksum=b1776092f2908d76e11fd6af87267469b2c17d5a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:45 managed-node1 python3.9[44186]: ansible-file Invoked with path=/home/user1/.config/containers/registries.conf.d state=directory owner=user1 group=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:46 managed-node1 python3.9[44335]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:52:46 managed-node1 python3.9[44455]: ansible-ansible.legacy.copy Invoked with dest=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf owner=user1 group=user1 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979165.7483666-13235-139178079209891/.source.conf _original_basename=.a1xet5pm follow=False checksum=fde25488ce7040f1639af7bfc88ed125318cc0b0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:46 managed-node1 python3.9[44604]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 group=user1 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:47 managed-node1 python3.9[44753]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:52:47 managed-node1 python3.9[44873]: ansible-ansible.legacy.copy Invoked with dest=/home/user1/.config/containers/storage.conf owner=user1 group=user1 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979166.9236264-13266-54800067456402/.source.conf _original_basename=.7t8_db15 follow=False checksum=38f015f4780579bd388dd955b42916199fd7fe19 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:47 managed-node1 python3.9[45022]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 group=user1 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:48 managed-node1 python3.9[45171]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:48 managed-node1 python3.9[45320]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:52:49 managed-node1 python3.9[45440]: ansible-ansible.legacy.copy Invoked with dest=/home/user1/.config/containers/policy.json owner=user1 group=user1 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979168.451584-13317-138138508589651/.source.json _original_basename=.tq0hx2zm follow=False checksum=6746c079ad563b735fc39f73d4876654b80b0a0d backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:49 managed-node1 python3.9[45589]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:50 managed-node1 python3.9[45740]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:50 managed-node1 python3.9[45890]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:51 managed-node1 python3.9[46040]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 group=user1 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:52:52 managed-node1 python3.9[46458]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:53 managed-node1 python3.9[46609]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:53 managed-node1 python3.9[46759]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:54 managed-node1 python3.9[46909]: ansible-stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:54 managed-node1 python3.9[47060]: ansible-stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:55 managed-node1 python3.9[47211]: ansible-stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:55 managed-node1 python3.9[47362]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:55 managed-node1 python3.9[47513]: ansible-ansible.legacy.command Invoked with _raw_params=grep 'container_name_as_hostname[ ]*=[ ]*true' /home/user1/.config/containers/containers.conf.d/50-systemroles.conf _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:57 managed-node1 python3.9[47812]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:58 managed-node1 python3.9[47967]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:52:59 managed-node1 python3.9[48118]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:52:59 managed-node1 python3.9[48268]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:53:00 managed-node1 python3.9[48418]: ansible-file Invoked with path=/home/user1/.config/containers/containers.conf.d state=directory owner=user1 group=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:00 managed-node1 python3.9[48567]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:01 managed-node1 python3.9[48642]: ansible-ansible.legacy.file Invoked with owner=user1 group=user1 mode=0644 dest=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf _original_basename=.os5cs4_x recurse=False state=file path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:01 managed-node1 python3.9[48791]: ansible-file Invoked with path=/home/user1/.config/containers/registries.conf.d state=directory owner=user1 group=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:02 managed-node1 python3.9[48940]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:02 managed-node1 python3.9[49015]: ansible-ansible.legacy.file Invoked with owner=user1 group=user1 mode=0644 dest=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf _original_basename=.9twz1wls recurse=False state=file path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:03 managed-node1 python3.9[49164]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 group=user1 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:03 managed-node1 python3.9[49313]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:03 managed-node1 python3.9[49388]: ansible-ansible.legacy.file Invoked with owner=user1 group=user1 mode=0644 dest=/home/user1/.config/containers/storage.conf _original_basename=.u72jitw8 recurse=False state=file path=/home/user1/.config/containers/storage.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:04 managed-node1 python3.9[49537]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 group=user1 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:04 managed-node1 python3.9[49686]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:04 managed-node1 python3.9[49837]: ansible-slurp Invoked with path=/home/user1/.config/containers/policy.json src=/home/user1/.config/containers/policy.json Jul 08 08:53:05 managed-node1 python3.9[49986]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:06 managed-node1 python3.9[50137]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:53:06 managed-node1 python3.9[50287]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:53:07 managed-node1 python3.9[50437]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 group=user1 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:09 managed-node1 python3.9[50810]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:09 managed-node1 python3.9[50961]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:53:10 managed-node1 python3.9[51111]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:53:10 managed-node1 python3.9[51261]: ansible-stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:11 managed-node1 python3.9[51412]: ansible-stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:11 managed-node1 python3.9[51563]: ansible-stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:12 managed-node1 python3.9[51714]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:13 managed-node1 python3.9[52014]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:53:14 managed-node1 python3.9[52169]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Jul 08 08:53:15 managed-node1 python3.9[52319]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:15 managed-node1 python3.9[52470]: ansible-file Invoked with path=/etc/containers/containers.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:16 managed-node1 python3.9[52619]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:16 managed-node1 python3.9[52739]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/containers.conf.d/50-systemroles.conf owner=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979196.068439-14128-196243318775352/.source.conf _original_basename=.9z01ulnk follow=False checksum=b1776092f2908d76e11fd6af87267469b2c17d5a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:17 managed-node1 python3.9[52888]: ansible-file Invoked with path=/etc/containers/registries.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:17 managed-node1 python3.9[53037]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:17 managed-node1 python3.9[53157]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/registries.conf.d/50-systemroles.conf owner=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979197.2308524-14157-252335483909182/.source.conf _original_basename=.wkt7vbqw follow=False checksum=fde25488ce7040f1639af7bfc88ed125318cc0b0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:18 managed-node1 python3.9[53306]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:18 managed-node1 python3.9[53455]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:18 managed-node1 python3.9[53577]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/storage.conf owner=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979198.3939052-14186-108211027079466/.source.conf _original_basename=.6_l0ebnm follow=False checksum=38f015f4780579bd388dd955b42916199fd7fe19 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:19 managed-node1 python3.9[53726]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:19 managed-node1 python3.9[53875]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:20 managed-node1 python3.9[54026]: ansible-slurp Invoked with path=/etc/containers/policy.json src=/etc/containers/policy.json Jul 08 08:53:20 managed-node1 python3.9[54175]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:21 managed-node1 python3.9[54297]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/policy.json owner=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979200.4464686-14254-119110580364058/.source.json _original_basename=.9grd0e5j follow=False checksum=6746c079ad563b735fc39f73d4876654b80b0a0d backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:21 managed-node1 python3.9[54446]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:22 managed-node1 python3.9[54597]: ansible-file Invoked with path=/root/.config/containers state=directory owner=root group=0 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:24 managed-node1 python3.9[55017]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:25 managed-node1 python3.9[55168]: ansible-stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:26 managed-node1 python3.9[55319]: ansible-stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:26 managed-node1 python3.9[55470]: ansible-stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:26 managed-node1 python3.9[55621]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:28 managed-node1 python3.9[55921]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:53:29 managed-node1 python3.9[56076]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:30 managed-node1 python3.9[56227]: ansible-file Invoked with path=/etc/containers/containers.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:30 managed-node1 python3.9[56376]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:31 managed-node1 python3.9[56451]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/containers.conf.d/50-systemroles.conf _original_basename=.lk6r9zg4 recurse=False state=file path=/etc/containers/containers.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:31 managed-node1 python3.9[56600]: ansible-file Invoked with path=/etc/containers/registries.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:32 managed-node1 python3.9[56749]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:32 managed-node1 python3.9[56824]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/registries.conf.d/50-systemroles.conf _original_basename=.0hqjoqag recurse=False state=file path=/etc/containers/registries.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:32 managed-node1 python3.9[56973]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:33 managed-node1 python3.9[57122]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:33 managed-node1 python3.9[57197]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/storage.conf _original_basename=.1ivx98ms recurse=False state=file path=/etc/containers/storage.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:33 managed-node1 python3.9[57346]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:34 managed-node1 python3.9[57495]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:34 managed-node1 python3.9[57646]: ansible-slurp Invoked with path=/etc/containers/policy.json src=/etc/containers/policy.json Jul 08 08:53:35 managed-node1 python3.9[57795]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:36 managed-node1 python3.9[57946]: ansible-file Invoked with path=/root/.config/containers state=directory owner=root group=0 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:37 managed-node1 python3.9[58319]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:38 managed-node1 python3.9[58470]: ansible-stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:39 managed-node1 python3.9[58621]: ansible-stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:39 managed-node1 python3.9[58772]: ansible-stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:40 managed-node1 python3.9[58923]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:40 managed-node1 python3.9[59074]: ansible-slurp Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf src=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf Jul 08 08:53:41 managed-node1 python3.9[59223]: ansible-slurp Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf src=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf Jul 08 08:53:41 managed-node1 python3.9[59372]: ansible-slurp Invoked with path=/home/user1/.config/containers/storage.conf src=/home/user1/.config/containers/storage.conf Jul 08 08:53:42 managed-node1 python3.9[59521]: ansible-slurp Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf src=/etc/containers/containers.conf.d/50-systemroles.conf Jul 08 08:53:42 managed-node1 python3.9[59670]: ansible-slurp Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf src=/etc/containers/registries.conf.d/50-systemroles.conf Jul 08 08:53:43 managed-node1 python3.9[59819]: ansible-slurp Invoked with path=/etc/containers/storage.conf src=/etc/containers/storage.conf Jul 08 08:53:45 managed-node1 python3.9[60117]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:53:46 managed-node1 python3.9[60273]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:47 managed-node1 python3.9[60424]: ansible-file Invoked with path=/etc/containers/containers.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:48 managed-node1 python3.9[60573]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:48 managed-node1 python3.9[60695]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/containers.conf.d/50-systemroles.conf owner=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979227.7911222-15133-177756245042776/.source.conf _original_basename=.eiu_kyxh follow=False checksum=9694c1d1c700a6435eecf4066b052584f4ee94c0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:48 managed-node1 python3.9[60844]: ansible-file Invoked with path=/etc/containers/registries.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:49 managed-node1 python3.9[60993]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:49 managed-node1 python3.9[61068]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/registries.conf.d/50-systemroles.conf _original_basename=.gr7jiie7 recurse=False state=file path=/etc/containers/registries.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:50 managed-node1 python3.9[61217]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:50 managed-node1 python3.9[61366]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:53:51 managed-node1 python3.9[61441]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/storage.conf _original_basename=.pcxrre0c recurse=False state=file path=/etc/containers/storage.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:51 managed-node1 python3.9[61590]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:52 managed-node1 python3.9[61739]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:53:52 managed-node1 python3.9[61890]: ansible-slurp Invoked with path=/etc/containers/policy.json src=/etc/containers/policy.json Jul 08 08:53:53 managed-node1 python3.9[62039]: ansible-slurp Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf src=/etc/containers/containers.conf.d/50-systemroles.conf Jul 08 08:53:53 managed-node1 python3.9[62188]: ansible-file Invoked with state=absent path=/etc/containers/containers.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:54 managed-node1 python3.9[62337]: ansible-file Invoked with state=absent path=/etc/containers/registries.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:54 managed-node1 python3.9[62486]: ansible-file Invoked with state=absent path=/etc/containers/storage.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:54 managed-node1 python3.9[62635]: ansible-file Invoked with state=absent path=/etc/containers/policy.json recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:55 managed-node1 python3.9[62784]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:55 managed-node1 python3.9[62933]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:56 managed-node1 python3.9[63082]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/storage.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:56 managed-node1 python3.9[63231]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/policy.json recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:56 managed-node1 python3.9[63380]: ansible-file Invoked with state=absent path=/root/.config/containers/auth.json recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:57 managed-node1 python3.9[63529]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/auth.json recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:53:57 managed-node1 python3.9[63678]: ansible-ansible.legacy.command Invoked with _raw_params=tar xfvpP /tmp/lsr_podman_config_6ihb6fh6/backup.tar _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:53:57 managed-node1 python3.9[63828]: ansible-file Invoked with state=absent path=/tmp/lsr_podman_config_6ihb6fh6 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:54:00 managed-node1 python3.9[64026]: ansible-setup Invoked with gather_subset=['!all', '!min', 'distribution', 'distribution_major_version', 'distribution_version', 'os_family'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Jul 08 08:54:00 managed-node1 python3.9[64177]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:01 managed-node1 python3.9[64326]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:03 managed-node1 python3.9[64624]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:54:03 managed-node1 python3.9[64779]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Jul 08 08:54:04 managed-node1 python3.9[64929]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:07 managed-node1 python3.9[65129]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Jul 08 08:54:10 managed-node1 python3.9[65304]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:11 managed-node1 python3.9[65453]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:13 managed-node1 python3.9[65751]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:54:14 managed-node1 python3.9[65907]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Jul 08 08:54:14 managed-node1 python3.9[66057]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:19 managed-node1 python3.9[66257]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Jul 08 08:54:20 managed-node1 python3.9[66432]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:21 managed-node1 python3.9[66581]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:23 managed-node1 python3.9[66879]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:54:23 managed-node1 python3.9[67034]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None Jul 08 08:54:24 managed-node1 python3.9[67184]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:27 managed-node1 python3.9[67335]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:28 managed-node1 python3.9[67486]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:54:29 managed-node1 python3.9[67635]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/nopull.container follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:54:29 managed-node1 python3.9[67755]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1751979269.1189396-16954-11122430624534/.source.container dest=/etc/containers/systemd/nopull.container owner=root group=0 mode=0644 follow=False _original_basename=systemd.j2 checksum=670d64fc68a9768edb20cad26df2acc703542d85 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:54:32 managed-node1 python3.9[68053]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:54:33 managed-node1 python3.9[68208]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:35 managed-node1 python3.9[68359]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:37 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 08 08:54:37 managed-node1 podman[68519]: 2025-07-08 08:54:37.140415246 -0400 EDT m=+0.021147061 image pull-error this_is_a_bogus_image:latest short-name resolution enforced but cannot prompt without a TTY Jul 08 08:54:37 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 08 08:54:37 managed-node1 python3.9[68674]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:54:38 managed-node1 python3.9[68823]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/bogus.container follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True Jul 08 08:54:38 managed-node1 python3.9[68943]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1751979277.771543-17272-2966124504745/.source.container dest=/etc/containers/systemd/bogus.container owner=root group=0 mode=0644 follow=False _original_basename=systemd.j2 checksum=1d087e679d135214e8ac9ccaf33b2222916efb7f backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Jul 08 08:54:41 managed-node1 python3.9[69241]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:54:42 managed-node1 python3.9[69396]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:45 managed-node1 python3.9[69547]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:46 managed-node1 python3.9[69698]: ansible-systemd Invoked with name=nopull.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None Jul 08 08:54:47 managed-node1 python3.9[69848]: ansible-stat Invoked with path=/etc/containers/systemd/nopull.container follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Jul 08 08:54:48 managed-node1 python3.9[70148]: ansible-ansible.legacy.command Invoked with _raw_params=set -x set -o pipefail exec 1>&2 #podman volume rm --all #podman network prune -f podman volume ls podman network ls podman secret ls podman container ls podman pod ls podman images systemctl list-units | grep quadlet _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:54:48 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 08 08:54:48 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 08 08:54:49 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully. â–‘â–‘ Subject: Unit succeeded â–‘â–‘ Defined-By: systemd â–‘â–‘ Support: https://access.redhat.com/support â–‘â–‘ â–‘â–‘ The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Jul 08 08:54:50 managed-node1 python3.9[70491]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Jul 08 08:54:51 managed-node1 python3.9[70646]: ansible-getent Invoked with database=passwd key=user_quadlet_basic fail_key=False service=None split=None Jul 08 08:54:51 managed-node1 python3.9[70796]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None PLAY RECAP ********************************************************************* managed-node1 : ok=110 changed=2 unreachable=0 failed=2 skipped=190 rescued=2 ignored=0 SYSTEM ROLES ERRORS BEGIN v1 [ { "ansible_version": "2.17.12", "end_time": "2025-07-08T12:54:47.976206+00:00Z", "host": "managed-node1", "message": "template error while templating string: Could not load \"podman_from_ini\": 'podman_from_ini'. String: {{ __podman_quadlet_raw.content | b64decode | podman_from_ini }}. Could not load \"podman_from_ini\": 'podman_from_ini'", "start_time": "2025-07-08T12:54:47.956781+00:00Z", "task_name": "Parse quadlet file", "task_path": "/tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:12" }, { "ansible_version": "2.17.12", "delta": "0:00:00.191455", "end_time": "2025-07-08 08:54:48.480918", "host": "managed-node1", "message": "non-zero return code", "rc": 1, "start_time": "2025-07-08 08:54:48.289463", "stderr": "+ set -o pipefail\n+ exec\n+ podman volume ls\nDRIVER VOLUME NAME\n+ podman network ls\nNETWORK ID NAME DRIVER\n2f259bab93aa podman bridge\n755dd383fce6 podman-default-kube-network bridge\n+ podman secret ls\nID NAME DRIVER CREATED UPDATED\n+ podman container ls\nCONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES\n+ podman pod ls\nPOD ID NAME STATUS CREATED INFRA ID # OF CONTAINERS\n+ podman images\nREPOSITORY TAG IMAGE ID CREATED SIZE\nquay.io/libpod/registry 2.8.2 0030ba3d620c 23 months ago 24.6 MB\nlocalhost:5000/libpod/testimage 20210610 9f9ec7f2fdef 4 years ago 7.99 MB\nquay.io/libpod/testimage 20210610 9f9ec7f2fdef 4 years ago 7.99 MB\n+ systemctl list-units\n+ grep quadlet", "task_name": "Debug3", "task_path": "/tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:270" }, { "ansible_version": "2.17.12", "end_time": "2025-07-08T12:54:51.199731+00:00Z", "host": "managed-node1", "message": "The given podman user [user_quadlet_basic] does not exist - cannot continue\n", "start_time": "2025-07-08T12:54:51.176833+00:00Z", "task_name": "Fail if user does not exist", "task_path": "/tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:9" }, { "ansible_version": "2.17.12", "delta": "0:00:00.052464", "end_time": "2025-07-08 08:54:51.567215", "host": "managed-node1", "message": "", "rc": 0, "start_time": "2025-07-08 08:54:51.514751", "stdout": "Jul 08 08:51:48 managed-node1 auth_test_1_kube-auth_test_1_kube[33421]: This container is intended for podman CI testing\nJul 08 08:51:48 managed-node1 podman[33425]: 2025-07-08 08:51:48.448578278 -0400 EDT m=+0.023969811 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:48 managed-node1 podman[33425]: 2025-07-08 08:51:48.463578624 -0400 EDT m=+0.038970410 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:48 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3715.\nJul 08 08:51:48 managed-node1 podman[33425]: 2025-07-08 08:51:48.513685618 -0400 EDT m=+0.089077175 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:48 managed-node1 podman[33425]: 2025-07-08 08:51:48.516933005 -0400 EDT m=+0.092324634 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:48 managed-node1 auth_test_1_kube-auth_test_1_kube[33460]: This container is intended for podman CI testing\nJul 08 08:51:48 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:48 managed-node1 podman[33464]: 2025-07-08 08:51:48.551624605 -0400 EDT m=+0.023398724 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:48 managed-node1 podman[33464]: 2025-07-08 08:51:48.564044867 -0400 EDT m=+0.035818947 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:48 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3720.\nJul 08 08:51:48 managed-node1 podman[33464]: 2025-07-08 08:51:48.610869632 -0400 EDT m=+0.082643798 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:48 managed-node1 podman[33464]: 2025-07-08 08:51:48.614112225 -0400 EDT m=+0.085886568 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:48 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:48 managed-node1 auth_test_1_kube-auth_test_1_kube[33475]: This container is intended for podman CI testing\nJul 08 08:51:48 managed-node1 podman[33479]: 2025-07-08 08:51:48.647583064 -0400 EDT m=+0.023183567 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:48 managed-node1 podman[33479]: 2025-07-08 08:51:48.661874275 -0400 EDT m=+0.037474756 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:48 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3725.\nJul 08 08:51:48 managed-node1 podman[33479]: 2025-07-08 08:51:48.712876218 -0400 EDT m=+0.088476803 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:48 managed-node1 podman[33479]: 2025-07-08 08:51:48.716219436 -0400 EDT m=+0.091820092 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:48 managed-node1 auth_test_1_kube-auth_test_1_kube[33490]: This container is intended for podman CI testing\nJul 08 08:51:48 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:48 managed-node1 conmon[33490]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:48 managed-node1 podman[33494]: 2025-07-08 08:51:48.751043262 -0400 EDT m=+0.023360199 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:48 managed-node1 podman[33494]: 2025-07-08 08:51:48.763434686 -0400 EDT m=+0.035751587 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:48 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3730.\nJul 08 08:51:48 managed-node1 podman[33494]: 2025-07-08 08:51:48.813391895 -0400 EDT m=+0.085708857 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:48 managed-node1 auth_test_1_kube-auth_test_1_kube[33506]: This container is intended for podman CI testing\nJul 08 08:51:48 managed-node1 conmon[33506]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:48 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:48 managed-node1 podman[33494]: 2025-07-08 08:51:48.817719676 -0400 EDT m=+0.090036659 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:48 managed-node1 podman[33510]: 2025-07-08 08:51:48.852289124 -0400 EDT m=+0.022895212 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:48 managed-node1 podman[33510]: 2025-07-08 08:51:48.864970582 -0400 EDT m=+0.035576641 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:48 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3735.\nJul 08 08:51:48 managed-node1 podman[33510]: 2025-07-08 08:51:48.914551976 -0400 EDT m=+0.085158053 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:48 managed-node1 podman[33510]: 2025-07-08 08:51:48.917640624 -0400 EDT m=+0.088246829 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:48 managed-node1 auth_test_1_kube-auth_test_1_kube[33521]: This container is intended for podman CI testing\nJul 08 08:51:48 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:48 managed-node1 podman[33525]: 2025-07-08 08:51:48.951310179 -0400 EDT m=+0.022752999 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:48 managed-node1 podman[33525]: 2025-07-08 08:51:48.963689665 -0400 EDT m=+0.035132460 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:48 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3740.\nJul 08 08:51:49 managed-node1 podman[33525]: 2025-07-08 08:51:49.008565374 -0400 EDT m=+0.080008215 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:49 managed-node1 podman[33525]: 2025-07-08 08:51:49.011684347 -0400 EDT m=+0.083127174 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33537]: This container is intended for podman CI testing\nJul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:49 managed-node1 podman[33541]: 2025-07-08 08:51:49.05246279 -0400 EDT m=+0.030949737 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:49 managed-node1 podman[33541]: 2025-07-08 08:51:49.067551179 -0400 EDT m=+0.046038114 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3745.\nJul 08 08:51:49 managed-node1 podman[33541]: 2025-07-08 08:51:49.126159103 -0400 EDT m=+0.104646383 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33598]: This container is intended for podman CI testing\nJul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:49 managed-node1 conmon[33598]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:49 managed-node1 podman[33541]: 2025-07-08 08:51:49.132373168 -0400 EDT m=+0.110860199 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:49 managed-node1 podman[33614]: 2025-07-08 08:51:49.17740571 -0400 EDT m=+0.030617289 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:49 managed-node1 podman[33614]: 2025-07-08 08:51:49.192793345 -0400 EDT m=+0.046004983 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3750.\nJul 08 08:51:49 managed-node1 podman[33614]: 2025-07-08 08:51:49.248278072 -0400 EDT m=+0.101490028 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33666]: This container is intended for podman CI testing\nJul 08 08:51:49 managed-node1 podman[33614]: 2025-07-08 08:51:49.253137795 -0400 EDT m=+0.106349379 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:49 managed-node1 podman[33689]: 2025-07-08 08:51:49.307534712 -0400 EDT m=+0.040092353 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:49 managed-node1 podman[33689]: 2025-07-08 08:51:49.330294602 -0400 EDT m=+0.062852173 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3755.\nJul 08 08:51:49 managed-node1 podman[33689]: 2025-07-08 08:51:49.417873351 -0400 EDT m=+0.150431000 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33707]: This container is intended for podman CI testing\nJul 08 08:51:49 managed-node1 conmon[33707]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:49 managed-node1 podman[33689]: 2025-07-08 08:51:49.425685791 -0400 EDT m=+0.158243377 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:49 managed-node1 python3.9[33701]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:51:49 managed-node1 podman[33711]: 2025-07-08 08:51:49.48629306 -0400 EDT m=+0.042105729 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:49 managed-node1 podman[33711]: 2025-07-08 08:51:49.505202864 -0400 EDT m=+0.061015306 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3760.\nJul 08 08:51:49 managed-node1 podman[33711]: 2025-07-08 08:51:49.563285914 -0400 EDT m=+0.119098412 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:49 managed-node1 podman[33711]: 2025-07-08 08:51:49.566298505 -0400 EDT m=+0.122111008 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33752]: This container is intended for podman CI testing\nJul 08 08:51:49 managed-node1 conmon[33752]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:49 managed-node1 podman[33756]: 2025-07-08 08:51:49.602293721 -0400 EDT m=+0.023104641 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:49 managed-node1 podman[33756]: 2025-07-08 08:51:49.61533105 -0400 EDT m=+0.036141951 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3765.\nJul 08 08:51:49 managed-node1 podman[33756]: 2025-07-08 08:51:49.660820556 -0400 EDT m=+0.081631519 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:49 managed-node1 podman[33756]: 2025-07-08 08:51:49.663846273 -0400 EDT m=+0.084657215 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33767]: This container is intended for podman CI testing\nJul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:49 managed-node1 conmon[33767]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:49 managed-node1 podman[33771]: 2025-07-08 08:51:49.696897085 -0400 EDT m=+0.021669345 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:49 managed-node1 podman[33771]: 2025-07-08 08:51:49.709430925 -0400 EDT m=+0.034203158 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3770.\nJul 08 08:51:49 managed-node1 rsyslogd[811]: imjournal: journal files changed, reloading... [v8.2412.0-2.el9 try https://www.rsyslog.com/e/0 ]\nJul 08 08:51:49 managed-node1 podman[33771]: 2025-07-08 08:51:49.762915854 -0400 EDT m=+0.087688178 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:49 managed-node1 podman[33771]: 2025-07-08 08:51:49.767703684 -0400 EDT m=+0.092476222 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33782]: This container is intended for podman CI testing\nJul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:49 managed-node1 podman[33787]: 2025-07-08 08:51:49.800414365 -0400 EDT m=+0.022705818 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:49 managed-node1 podman[33787]: 2025-07-08 08:51:49.813402246 -0400 EDT m=+0.035693631 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3775.\nJul 08 08:51:49 managed-node1 podman[33787]: 2025-07-08 08:51:49.859216266 -0400 EDT m=+0.081507708 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:49 managed-node1 podman[33787]: 2025-07-08 08:51:49.862342028 -0400 EDT m=+0.084633601 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33798]: This container is intended for podman CI testing\nJul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:49 managed-node1 podman[33802]: 2025-07-08 08:51:49.894773177 -0400 EDT m=+0.022994496 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:49 managed-node1 podman[33802]: 2025-07-08 08:51:49.907297159 -0400 EDT m=+0.035518448 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:49 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3780.\nJul 08 08:51:49 managed-node1 podman[33802]: 2025-07-08 08:51:49.954555528 -0400 EDT m=+0.082776810 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:49 managed-node1 podman[33802]: 2025-07-08 08:51:49.95854713 -0400 EDT m=+0.086768463 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:49 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:49 managed-node1 auth_test_1_kube-auth_test_1_kube[33813]: This container is intended for podman CI testing\nJul 08 08:51:49 managed-node1 conmon[33813]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:49 managed-node1 podman[33817]: 2025-07-08 08:51:49.994321706 -0400 EDT m=+0.021682296 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:50 managed-node1 podman[33817]: 2025-07-08 08:51:50.007308802 -0400 EDT m=+0.034669150 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3785.\nJul 08 08:51:50 managed-node1 podman[33817]: 2025-07-08 08:51:50.06483917 -0400 EDT m=+0.092199738 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:50 managed-node1 podman[33817]: 2025-07-08 08:51:50.069602607 -0400 EDT m=+0.096962923 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[33828]: This container is intended for podman CI testing\nJul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:50 managed-node1 conmon[33828]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:50 managed-node1 podman[33854]: 2025-07-08 08:51:50.124935034 -0400 EDT m=+0.038208882 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:50 managed-node1 podman[33854]: 2025-07-08 08:51:50.13785611 -0400 EDT m=+0.051130005 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3790.\nJul 08 08:51:50 managed-node1 podman[33854]: 2025-07-08 08:51:50.202569983 -0400 EDT m=+0.115844060 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:50 managed-node1 podman[33854]: 2025-07-08 08:51:50.20647275 -0400 EDT m=+0.119746611 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[33918]: This container is intended for podman CI testing\nJul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:50 managed-node1 podman[33940]: 2025-07-08 08:51:50.255118585 -0400 EDT m=+0.033291310 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:50 managed-node1 podman[33940]: 2025-07-08 08:51:50.26846983 -0400 EDT m=+0.046642515 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3795.\nJul 08 08:51:50 managed-node1 podman[33940]: 2025-07-08 08:51:50.369192726 -0400 EDT m=+0.147365515 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[33983]: This container is intended for podman CI testing\nJul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:50 managed-node1 conmon[33983]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:50 managed-node1 podman[33940]: 2025-07-08 08:51:50.376002241 -0400 EDT m=+0.154174886 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:50 managed-node1 python3.9[33978]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:51:50 managed-node1 podman[33987]: 2025-07-08 08:51:50.4323887 -0400 EDT m=+0.039929508 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:50 managed-node1 podman[33987]: 2025-07-08 08:51:50.446216676 -0400 EDT m=+0.053757208 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3800.\nJul 08 08:51:50 managed-node1 podman[33987]: 2025-07-08 08:51:50.504432318 -0400 EDT m=+0.111972903 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:50 managed-node1 podman[33987]: 2025-07-08 08:51:50.507589487 -0400 EDT m=+0.115130423 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[34024]: This container is intended for podman CI testing\nJul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:50 managed-node1 podman[34028]: 2025-07-08 08:51:50.543336091 -0400 EDT m=+0.023133313 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:50 managed-node1 podman[34028]: 2025-07-08 08:51:50.557285225 -0400 EDT m=+0.037082354 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3805.\nJul 08 08:51:50 managed-node1 podman[34028]: 2025-07-08 08:51:50.607148898 -0400 EDT m=+0.086946107 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:50 managed-node1 podman[34028]: 2025-07-08 08:51:50.610448275 -0400 EDT m=+0.090245452 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[34039]: This container is intended for podman CI testing\nJul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:50 managed-node1 podman[34043]: 2025-07-08 08:51:50.64215535 -0400 EDT m=+0.022131003 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:50 managed-node1 podman[34043]: 2025-07-08 08:51:50.655139945 -0400 EDT m=+0.035115543 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3810.\nJul 08 08:51:50 managed-node1 podman[34043]: 2025-07-08 08:51:50.706854723 -0400 EDT m=+0.086830396 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:50 managed-node1 podman[34043]: 2025-07-08 08:51:50.710102295 -0400 EDT m=+0.090077940 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[34054]: This container is intended for podman CI testing\nJul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:50 managed-node1 podman[34058]: 2025-07-08 08:51:50.746221367 -0400 EDT m=+0.022863699 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:50 managed-node1 podman[34058]: 2025-07-08 08:51:50.758828206 -0400 EDT m=+0.035470533 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3815.\nJul 08 08:51:50 managed-node1 podman[34058]: 2025-07-08 08:51:50.808790018 -0400 EDT m=+0.085432443 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[34069]: This container is intended for podman CI testing\nJul 08 08:51:50 managed-node1 conmon[34069]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:50 managed-node1 podman[34058]: 2025-07-08 08:51:50.81308862 -0400 EDT m=+0.089730948 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:50 managed-node1 podman[34073]: 2025-07-08 08:51:50.846088349 -0400 EDT m=+0.022900304 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:50 managed-node1 podman[34073]: 2025-07-08 08:51:50.858864628 -0400 EDT m=+0.035676518 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3820.\nJul 08 08:51:50 managed-node1 podman[34073]: 2025-07-08 08:51:50.907379988 -0400 EDT m=+0.084191923 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:50 managed-node1 podman[34073]: 2025-07-08 08:51:50.910733502 -0400 EDT m=+0.087545474 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:50 managed-node1 auth_test_1_kube-auth_test_1_kube[34084]: This container is intended for podman CI testing\nJul 08 08:51:50 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:50 managed-node1 podman[34088]: 2025-07-08 08:51:50.946679837 -0400 EDT m=+0.022865188 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:50 managed-node1 podman[34088]: 2025-07-08 08:51:50.959995671 -0400 EDT m=+0.036181000 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:50 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3825.\nJul 08 08:51:51 managed-node1 podman[34088]: 2025-07-08 08:51:51.002881906 -0400 EDT m=+0.079067276 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:51 managed-node1 podman[34088]: 2025-07-08 08:51:51.006032573 -0400 EDT m=+0.082218026 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34099]: This container is intended for podman CI testing\nJul 08 08:51:51 managed-node1 podman[34103]: 2025-07-08 08:51:51.041650332 -0400 EDT m=+0.022409183 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:51 managed-node1 podman[34103]: 2025-07-08 08:51:51.054826908 -0400 EDT m=+0.035585744 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3830.\nJul 08 08:51:51 managed-node1 podman[34103]: 2025-07-08 08:51:51.10523635 -0400 EDT m=+0.085995171 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:51 managed-node1 podman[34103]: 2025-07-08 08:51:51.108479715 -0400 EDT m=+0.089238666 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34114]: This container is intended for podman CI testing\nJul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:51 managed-node1 conmon[34114]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:51 managed-node1 podman[34118]: 2025-07-08 08:51:51.145983784 -0400 EDT m=+0.023131545 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:51 managed-node1 podman[34118]: 2025-07-08 08:51:51.158714763 -0400 EDT m=+0.035862464 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3835.\nJul 08 08:51:51 managed-node1 podman[34118]: 2025-07-08 08:51:51.206515345 -0400 EDT m=+0.083663128 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:51 managed-node1 podman[34118]: 2025-07-08 08:51:51.209753192 -0400 EDT m=+0.086900961 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34129]: This container is intended for podman CI testing\nJul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:51 managed-node1 conmon[34129]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:51 managed-node1 podman[34133]: 2025-07-08 08:51:51.246626352 -0400 EDT m=+0.023308457 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:51 managed-node1 podman[34133]: 2025-07-08 08:51:51.259550155 -0400 EDT m=+0.036232229 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3840.\nJul 08 08:51:51 managed-node1 podman[34133]: 2025-07-08 08:51:51.310043813 -0400 EDT m=+0.086725970 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:51 managed-node1 podman[34133]: 2025-07-08 08:51:51.314825939 -0400 EDT m=+0.091508123 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34144]: This container is intended for podman CI testing\nJul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:51 managed-node1 conmon[34144]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:51 managed-node1 podman[34148]: 2025-07-08 08:51:51.349444872 -0400 EDT m=+0.020834015 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:51 managed-node1 podman[34148]: 2025-07-08 08:51:51.362030702 -0400 EDT m=+0.033419755 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3845.\nJul 08 08:51:51 managed-node1 podman[34148]: 2025-07-08 08:51:51.411294271 -0400 EDT m=+0.082683322 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:51 managed-node1 podman[34148]: 2025-07-08 08:51:51.414607833 -0400 EDT m=+0.085996969 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34159]: This container is intended for podman CI testing\nJul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:51 managed-node1 podman[34163]: 2025-07-08 08:51:51.451910069 -0400 EDT m=+0.024032297 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:51 managed-node1 podman[34163]: 2025-07-08 08:51:51.465095028 -0400 EDT m=+0.037217295 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3850.\nJul 08 08:51:51 managed-node1 podman[34163]: 2025-07-08 08:51:51.51842398 -0400 EDT m=+0.090546232 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:51 managed-node1 podman[34163]: 2025-07-08 08:51:51.521671054 -0400 EDT m=+0.093793300 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34174]: This container is intended for podman CI testing\nJul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:51 managed-node1 podman[34178]: 2025-07-08 08:51:51.557171448 -0400 EDT m=+0.026340600 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:51 managed-node1 podman[34178]: 2025-07-08 08:51:51.570270377 -0400 EDT m=+0.039439505 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3855.\nJul 08 08:51:51 managed-node1 podman[34178]: 2025-07-08 08:51:51.623425288 -0400 EDT m=+0.092594547 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:51 managed-node1 podman[34178]: 2025-07-08 08:51:51.626725482 -0400 EDT m=+0.095894645 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34190]: This container is intended for podman CI testing\nJul 08 08:51:51 managed-node1 podman[34194]: 2025-07-08 08:51:51.66428524 -0400 EDT m=+0.023796852 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:51 managed-node1 podman[34194]: 2025-07-08 08:51:51.677129052 -0400 EDT m=+0.036640614 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3860.\nJul 08 08:51:51 managed-node1 podman[34194]: 2025-07-08 08:51:51.728717704 -0400 EDT m=+0.088229325 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:51 managed-node1 podman[34194]: 2025-07-08 08:51:51.731914334 -0400 EDT m=+0.091425975 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34205]: This container is intended for podman CI testing\nJul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:51 managed-node1 conmon[34205]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:51 managed-node1 podman[34209]: 2025-07-08 08:51:51.766271597 -0400 EDT m=+0.022524310 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:51 managed-node1 podman[34209]: 2025-07-08 08:51:51.778901663 -0400 EDT m=+0.035154314 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3865.\nJul 08 08:51:51 managed-node1 podman[34209]: 2025-07-08 08:51:51.827854907 -0400 EDT m=+0.084107688 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34220]: This container is intended for podman CI testing\nJul 08 08:51:51 managed-node1 podman[34209]: 2025-07-08 08:51:51.831526913 -0400 EDT m=+0.087779765 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:51 managed-node1 conmon[34220]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:51 managed-node1 podman[34224]: 2025-07-08 08:51:51.868178779 -0400 EDT m=+0.023709243 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:51 managed-node1 podman[34224]: 2025-07-08 08:51:51.881047661 -0400 EDT m=+0.036578030 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:51 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3870.\nJul 08 08:51:51 managed-node1 podman[34224]: 2025-07-08 08:51:51.93290684 -0400 EDT m=+0.088437271 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:51 managed-node1 podman[34224]: 2025-07-08 08:51:51.936102953 -0400 EDT m=+0.091633364 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:51 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:51 managed-node1 auth_test_1_kube-auth_test_1_kube[34235]: This container is intended for podman CI testing\nJul 08 08:51:51 managed-node1 conmon[34235]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:51 managed-node1 podman[34239]: 2025-07-08 08:51:51.974911564 -0400 EDT m=+0.023383080 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:51 managed-node1 podman[34239]: 2025-07-08 08:51:51.98759638 -0400 EDT m=+0.036067772 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3875.\nJul 08 08:51:52 managed-node1 podman[34239]: 2025-07-08 08:51:52.037897828 -0400 EDT m=+0.086369361 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:52 managed-node1 podman[34239]: 2025-07-08 08:51:52.04121166 -0400 EDT m=+0.089683114 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34250]: This container is intended for podman CI testing\nJul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:52 managed-node1 podman[34254]: 2025-07-08 08:51:52.076459043 -0400 EDT m=+0.022091035 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:52 managed-node1 podman[34254]: 2025-07-08 08:51:52.089098566 -0400 EDT m=+0.034730678 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3880.\nJul 08 08:51:52 managed-node1 podman[34254]: 2025-07-08 08:51:52.136605391 -0400 EDT m=+0.082237394 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:52 managed-node1 podman[34254]: 2025-07-08 08:51:52.139726463 -0400 EDT m=+0.085358496 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34266]: This container is intended for podman CI testing\nJul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:52 managed-node1 conmon[34266]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:52 managed-node1 podman[34270]: 2025-07-08 08:51:52.177785317 -0400 EDT m=+0.024693442 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:52 managed-node1 podman[34270]: 2025-07-08 08:51:52.190676452 -0400 EDT m=+0.037584544 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3885.\nJul 08 08:51:52 managed-node1 podman[34270]: 2025-07-08 08:51:52.242920332 -0400 EDT m=+0.089828489 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34281]: This container is intended for podman CI testing\nJul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:52 managed-node1 podman[34270]: 2025-07-08 08:51:52.247101011 -0400 EDT m=+0.094009235 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:52 managed-node1 conmon[34281]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:52 managed-node1 podman[34285]: 2025-07-08 08:51:52.281399672 -0400 EDT m=+0.024927817 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:52 managed-node1 podman[34285]: 2025-07-08 08:51:52.299134966 -0400 EDT m=+0.042663101 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3890.\nJul 08 08:51:52 managed-node1 podman[34285]: 2025-07-08 08:51:52.343500246 -0400 EDT m=+0.087028594 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:52 managed-node1 podman[34285]: 2025-07-08 08:51:52.349245849 -0400 EDT m=+0.092773962 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34296]: This container is intended for podman CI testing\nJul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:52 managed-node1 conmon[34296]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:52 managed-node1 podman[34300]: 2025-07-08 08:51:52.387822334 -0400 EDT m=+0.024037871 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:52 managed-node1 podman[34300]: 2025-07-08 08:51:52.400754882 -0400 EDT m=+0.036970368 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3895.\nJul 08 08:51:52 managed-node1 podman[34300]: 2025-07-08 08:51:52.45095469 -0400 EDT m=+0.087170281 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:52 managed-node1 podman[34300]: 2025-07-08 08:51:52.454642211 -0400 EDT m=+0.090857887 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34312]: This container is intended for podman CI testing\nJul 08 08:51:52 managed-node1 conmon[34312]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:52 managed-node1 podman[34316]: 2025-07-08 08:51:52.489368338 -0400 EDT m=+0.023100495 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:52 managed-node1 podman[34316]: 2025-07-08 08:51:52.501783549 -0400 EDT m=+0.035515669 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3900.\nJul 08 08:51:52 managed-node1 podman[34316]: 2025-07-08 08:51:52.550546694 -0400 EDT m=+0.084278835 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:52 managed-node1 podman[34316]: 2025-07-08 08:51:52.553615063 -0400 EDT m=+0.087347258 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34327]: This container is intended for podman CI testing\nJul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:52 managed-node1 conmon[34327]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:52 managed-node1 podman[34331]: 2025-07-08 08:51:52.595944229 -0400 EDT m=+0.032198277 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:52 managed-node1 podman[34331]: 2025-07-08 08:51:52.611282013 -0400 EDT m=+0.047535952 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3905.\nJul 08 08:51:52 managed-node1 podman[34331]: 2025-07-08 08:51:52.669145562 -0400 EDT m=+0.105399639 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34392]: This container is intended for podman CI testing\nJul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:52 managed-node1 podman[34331]: 2025-07-08 08:51:52.676901314 -0400 EDT m=+0.113155204 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:52 managed-node1 podman[34411]: 2025-07-08 08:51:52.717264706 -0400 EDT m=+0.027508917 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:52 managed-node1 podman[34411]: 2025-07-08 08:51:52.730524627 -0400 EDT m=+0.040768929 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3910.\nJul 08 08:51:52 managed-node1 podman[34411]: 2025-07-08 08:51:52.786375883 -0400 EDT m=+0.096620068 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:52 managed-node1 podman[34411]: 2025-07-08 08:51:52.790202655 -0400 EDT m=+0.100446862 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34456]: This container is intended for podman CI testing\nJul 08 08:51:52 managed-node1 podman[34486]: 2025-07-08 08:51:52.84544824 -0400 EDT m=+0.036617469 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:52 managed-node1 podman[34486]: 2025-07-08 08:51:52.859497057 -0400 EDT m=+0.050666352 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:52 managed-node1 python3.9[34485]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:51:52 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3915.\nJul 08 08:51:52 managed-node1 podman[34486]: 2025-07-08 08:51:52.968714976 -0400 EDT m=+0.159884342 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:52 managed-node1 auth_test_1_kube-auth_test_1_kube[34498]: This container is intended for podman CI testing\nJul 08 08:51:52 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:52 managed-node1 podman[34486]: 2025-07-08 08:51:52.97456631 -0400 EDT m=+0.165735868 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:53 managed-node1 podman[34504]: 2025-07-08 08:51:53.023986337 -0400 EDT m=+0.037748085 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:53 managed-node1 podman[34504]: 2025-07-08 08:51:53.037392078 -0400 EDT m=+0.051153917 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3920.\nJul 08 08:51:53 managed-node1 podman[34504]: 2025-07-08 08:51:53.086015594 -0400 EDT m=+0.099777537 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:53 managed-node1 podman[34504]: 2025-07-08 08:51:53.090468412 -0400 EDT m=+0.104230229 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34539]: This container is intended for podman CI testing\nJul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:53 managed-node1 conmon[34539]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:53 managed-node1 podman[34543]: 2025-07-08 08:51:53.124650233 -0400 EDT m=+0.021518634 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:53 managed-node1 podman[34543]: 2025-07-08 08:51:53.137900177 -0400 EDT m=+0.034768685 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3925.\nJul 08 08:51:53 managed-node1 podman[34543]: 2025-07-08 08:51:53.189829619 -0400 EDT m=+0.086698081 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:53 managed-node1 podman[34543]: 2025-07-08 08:51:53.193012125 -0400 EDT m=+0.089880597 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34555]: This container is intended for podman CI testing\nJul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:53 managed-node1 podman[34559]: 2025-07-08 08:51:53.22851641 -0400 EDT m=+0.026156436 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:53 managed-node1 podman[34559]: 2025-07-08 08:51:53.241177765 -0400 EDT m=+0.038817802 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3930.\nJul 08 08:51:53 managed-node1 podman[34559]: 2025-07-08 08:51:53.28736722 -0400 EDT m=+0.085007417 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:53 managed-node1 podman[34559]: 2025-07-08 08:51:53.292001399 -0400 EDT m=+0.089641461 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34570]: This container is intended for podman CI testing\nJul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:53 managed-node1 podman[34574]: 2025-07-08 08:51:53.326616039 -0400 EDT m=+0.022763830 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:53 managed-node1 podman[34574]: 2025-07-08 08:51:53.339468171 -0400 EDT m=+0.035615893 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3935.\nJul 08 08:51:53 managed-node1 podman[34574]: 2025-07-08 08:51:53.381558322 -0400 EDT m=+0.077706147 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:53 managed-node1 podman[34574]: 2025-07-08 08:51:53.386169788 -0400 EDT m=+0.082317612 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34585]: This container is intended for podman CI testing\nJul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:53 managed-node1 podman[34589]: 2025-07-08 08:51:53.418087203 -0400 EDT m=+0.022395652 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:53 managed-node1 podman[34589]: 2025-07-08 08:51:53.431253604 -0400 EDT m=+0.035562005 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3940.\nJul 08 08:51:53 managed-node1 podman[34589]: 2025-07-08 08:51:53.477797896 -0400 EDT m=+0.082106459 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:53 managed-node1 podman[34589]: 2025-07-08 08:51:53.480897464 -0400 EDT m=+0.085205877 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34600]: This container is intended for podman CI testing\nJul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:53 managed-node1 podman[34604]: 2025-07-08 08:51:53.517451555 -0400 EDT m=+0.023458453 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:53 managed-node1 podman[34604]: 2025-07-08 08:51:53.530275652 -0400 EDT m=+0.036282517 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3945.\nJul 08 08:51:53 managed-node1 podman[34604]: 2025-07-08 08:51:53.575783768 -0400 EDT m=+0.081790707 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:53 managed-node1 podman[34604]: 2025-07-08 08:51:53.580302043 -0400 EDT m=+0.086308990 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34615]: This container is intended for podman CI testing\nJul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:53 managed-node1 conmon[34615]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:53 managed-node1 podman[34619]: 2025-07-08 08:51:53.616636269 -0400 EDT m=+0.023020507 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:53 managed-node1 podman[34619]: 2025-07-08 08:51:53.629586441 -0400 EDT m=+0.035970587 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3950.\nJul 08 08:51:53 managed-node1 podman[34619]: 2025-07-08 08:51:53.678750449 -0400 EDT m=+0.085134654 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:53 managed-node1 podman[34619]: 2025-07-08 08:51:53.681880867 -0400 EDT m=+0.088265167 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34630]: This container is intended for podman CI testing\nJul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:53 managed-node1 conmon[34630]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:53 managed-node1 podman[34634]: 2025-07-08 08:51:53.717940888 -0400 EDT m=+0.022864499 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:53 managed-node1 podman[34634]: 2025-07-08 08:51:53.730421158 -0400 EDT m=+0.035344744 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3955.\nJul 08 08:51:53 managed-node1 podman[34634]: 2025-07-08 08:51:53.779644519 -0400 EDT m=+0.084568160 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34645]: This container is intended for podman CI testing\nJul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:53 managed-node1 podman[34634]: 2025-07-08 08:51:53.785388726 -0400 EDT m=+0.090312439 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:53 managed-node1 conmon[34645]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:53 managed-node1 podman[34649]: 2025-07-08 08:51:53.818900307 -0400 EDT m=+0.024214971 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:53 managed-node1 podman[34649]: 2025-07-08 08:51:53.83171962 -0400 EDT m=+0.037034251 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3960.\nJul 08 08:51:53 managed-node1 podman[34649]: 2025-07-08 08:51:53.884257137 -0400 EDT m=+0.089571850 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:53 managed-node1 podman[34649]: 2025-07-08 08:51:53.888055587 -0400 EDT m=+0.093370276 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34661]: This container is intended for podman CI testing\nJul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:53 managed-node1 podman[34665]: 2025-07-08 08:51:53.920914098 -0400 EDT m=+0.021177869 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:53 managed-node1 podman[34665]: 2025-07-08 08:51:53.935462662 -0400 EDT m=+0.035726391 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:53 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3965.\nJul 08 08:51:53 managed-node1 podman[34665]: 2025-07-08 08:51:53.984659847 -0400 EDT m=+0.084923716 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:53 managed-node1 podman[34665]: 2025-07-08 08:51:53.987748479 -0400 EDT m=+0.088012382 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:53 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:53 managed-node1 auth_test_1_kube-auth_test_1_kube[34676]: This container is intended for podman CI testing\nJul 08 08:51:54 managed-node1 podman[34680]: 2025-07-08 08:51:54.023217593 -0400 EDT m=+0.022450999 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:54 managed-node1 podman[34680]: 2025-07-08 08:51:54.035905475 -0400 EDT m=+0.035138812 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3970.\nJul 08 08:51:54 managed-node1 podman[34680]: 2025-07-08 08:51:54.085861249 -0400 EDT m=+0.085094690 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:54 managed-node1 podman[34680]: 2025-07-08 08:51:54.08899507 -0400 EDT m=+0.088228367 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34692]: This container is intended for podman CI testing\nJul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:54 managed-node1 conmon[34692]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:54 managed-node1 podman[34696]: 2025-07-08 08:51:54.124730926 -0400 EDT m=+0.025816549 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:54 managed-node1 podman[34696]: 2025-07-08 08:51:54.137552729 -0400 EDT m=+0.038638342 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3975.\nJul 08 08:51:54 managed-node1 podman[34696]: 2025-07-08 08:51:54.183060683 -0400 EDT m=+0.084146331 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34707]: This container is intended for podman CI testing\nJul 08 08:51:54 managed-node1 podman[34696]: 2025-07-08 08:51:54.186592319 -0400 EDT m=+0.087677967 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:54 managed-node1 conmon[34707]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:54 managed-node1 podman[34711]: 2025-07-08 08:51:54.222780153 -0400 EDT m=+0.023561311 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:54 managed-node1 podman[34711]: 2025-07-08 08:51:54.235623964 -0400 EDT m=+0.036405063 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3980.\nJul 08 08:51:54 managed-node1 podman[34711]: 2025-07-08 08:51:54.284151982 -0400 EDT m=+0.084933170 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34723]: This container is intended for podman CI testing\nJul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:54 managed-node1 conmon[34723]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:54 managed-node1 podman[34711]: 2025-07-08 08:51:54.290397046 -0400 EDT m=+0.091178256 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:54 managed-node1 podman[34727]: 2025-07-08 08:51:54.321959858 -0400 EDT m=+0.023134363 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:54 managed-node1 podman[34727]: 2025-07-08 08:51:54.334452722 -0400 EDT m=+0.035627222 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3985.\nJul 08 08:51:54 managed-node1 podman[34727]: 2025-07-08 08:51:54.382188683 -0400 EDT m=+0.083363410 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34738]: This container is intended for podman CI testing\nJul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:54 managed-node1 conmon[34738]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:54 managed-node1 podman[34727]: 2025-07-08 08:51:54.389123428 -0400 EDT m=+0.090297921 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:54 managed-node1 podman[34742]: 2025-07-08 08:51:54.421808862 -0400 EDT m=+0.022864231 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:54 managed-node1 podman[34742]: 2025-07-08 08:51:54.434499651 -0400 EDT m=+0.035554851 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3990.\nJul 08 08:51:54 managed-node1 podman[34742]: 2025-07-08 08:51:54.482427396 -0400 EDT m=+0.083482615 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:54 managed-node1 podman[34742]: 2025-07-08 08:51:54.485442521 -0400 EDT m=+0.086497760 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34753]: This container is intended for podman CI testing\nJul 08 08:51:54 managed-node1 podman[34757]: 2025-07-08 08:51:54.519902654 -0400 EDT m=+0.021765549 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:54 managed-node1 podman[34757]: 2025-07-08 08:51:54.532609106 -0400 EDT m=+0.034471955 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 3995.\nJul 08 08:51:54 managed-node1 podman[34757]: 2025-07-08 08:51:54.584264438 -0400 EDT m=+0.086127398 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:54 managed-node1 podman[34757]: 2025-07-08 08:51:54.587333117 -0400 EDT m=+0.089196094 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34768]: This container is intended for podman CI testing\nJul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:54 managed-node1 conmon[34768]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:54 managed-node1 podman[34772]: 2025-07-08 08:51:54.622454787 -0400 EDT m=+0.022629409 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:54 managed-node1 podman[34772]: 2025-07-08 08:51:54.635082693 -0400 EDT m=+0.035257419 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4000.\nJul 08 08:51:54 managed-node1 podman[34772]: 2025-07-08 08:51:54.68244541 -0400 EDT m=+0.082620046 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:54 managed-node1 podman[34772]: 2025-07-08 08:51:54.685503704 -0400 EDT m=+0.085678387 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34784]: This container is intended for podman CI testing\nJul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:54 managed-node1 conmon[34784]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:54 managed-node1 podman[34788]: 2025-07-08 08:51:54.720325287 -0400 EDT m=+0.023012889 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:54 managed-node1 podman[34788]: 2025-07-08 08:51:54.733130807 -0400 EDT m=+0.035818363 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4005.\nJul 08 08:51:54 managed-node1 podman[34788]: 2025-07-08 08:51:54.781809086 -0400 EDT m=+0.084497114 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:54 managed-node1 podman[34788]: 2025-07-08 08:51:54.784961617 -0400 EDT m=+0.087649271 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34799]: This container is intended for podman CI testing\nJul 08 08:51:54 managed-node1 conmon[34799]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:54 managed-node1 podman[34803]: 2025-07-08 08:51:54.823925637 -0400 EDT m=+0.025199186 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:54 managed-node1 podman[34803]: 2025-07-08 08:51:54.837754264 -0400 EDT m=+0.039027530 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:54 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4010.\nJul 08 08:51:54 managed-node1 podman[34803]: 2025-07-08 08:51:54.898369568 -0400 EDT m=+0.099642684 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:54 managed-node1 auth_test_1_kube-auth_test_1_kube[34846]: This container is intended for podman CI testing\nJul 08 08:51:54 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:54 managed-node1 podman[34803]: 2025-07-08 08:51:54.90352077 -0400 EDT m=+0.104793920 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:54 managed-node1 conmon[34846]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:54 managed-node1 podman[34870]: 2025-07-08 08:51:54.956835937 -0400 EDT m=+0.043874313 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:54 managed-node1 podman[34870]: 2025-07-08 08:51:54.970262097 -0400 EDT m=+0.057300532 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4015.\nJul 08 08:51:55 managed-node1 podman[34870]: 2025-07-08 08:51:55.030784056 -0400 EDT m=+0.117822460 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:55 managed-node1 podman[34870]: 2025-07-08 08:51:55.034551669 -0400 EDT m=+0.121590030 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:55 managed-node1 conmon[34928]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[34928]: This container is intended for podman CI testing\nJul 08 08:51:55 managed-node1 podman[34949]: 2025-07-08 08:51:55.083570411 -0400 EDT m=+0.033836426 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:55 managed-node1 podman[34949]: 2025-07-08 08:51:55.101627249 -0400 EDT m=+0.051893114 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4020.\nJul 08 08:51:55 managed-node1 podman[34949]: 2025-07-08 08:51:55.162314773 -0400 EDT m=+0.112580805 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[34969]: This container is intended for podman CI testing\nJul 08 08:51:55 managed-node1 conmon[34969]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:55 managed-node1 podman[34949]: 2025-07-08 08:51:55.169334312 -0400 EDT m=+0.119600323 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:55 managed-node1 podman[34973]: 2025-07-08 08:51:55.2233046 -0400 EDT m=+0.039432681 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:55 managed-node1 python3.9[34962]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:51:55 managed-node1 podman[34973]: 2025-07-08 08:51:55.241773549 -0400 EDT m=+0.057901492 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4025.\nJul 08 08:51:55 managed-node1 podman[34973]: 2025-07-08 08:51:55.298195371 -0400 EDT m=+0.114323417 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:55 managed-node1 podman[34973]: 2025-07-08 08:51:55.302460802 -0400 EDT m=+0.118588843 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[34999]: This container is intended for podman CI testing\nJul 08 08:51:55 managed-node1 conmon[34999]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:55 managed-node1 podman[35015]: 2025-07-08 08:51:55.33487409 -0400 EDT m=+0.022966278 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:55 managed-node1 podman[35015]: 2025-07-08 08:51:55.347928004 -0400 EDT m=+0.036020177 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4030.\nJul 08 08:51:55 managed-node1 podman[35015]: 2025-07-08 08:51:55.39641325 -0400 EDT m=+0.084505554 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:55 managed-node1 podman[35015]: 2025-07-08 08:51:55.399632598 -0400 EDT m=+0.087724869 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35026]: This container is intended for podman CI testing\nJul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:55 managed-node1 conmon[35026]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:55 managed-node1 podman[35030]: 2025-07-08 08:51:55.432369825 -0400 EDT m=+0.021888850 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:55 managed-node1 podman[35030]: 2025-07-08 08:51:55.444847265 -0400 EDT m=+0.034366252 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4035.\nJul 08 08:51:55 managed-node1 podman[35030]: 2025-07-08 08:51:55.494822879 -0400 EDT m=+0.084342000 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:55 managed-node1 podman[35030]: 2025-07-08 08:51:55.49802887 -0400 EDT m=+0.087547889 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35041]: This container is intended for podman CI testing\nJul 08 08:51:55 managed-node1 conmon[35041]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:55 managed-node1 podman[35045]: 2025-07-08 08:51:55.534978267 -0400 EDT m=+0.022300939 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:55 managed-node1 podman[35045]: 2025-07-08 08:51:55.547819511 -0400 EDT m=+0.035142129 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4040.\nJul 08 08:51:55 managed-node1 podman[35045]: 2025-07-08 08:51:55.596409489 -0400 EDT m=+0.083732130 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:55 managed-node1 podman[35045]: 2025-07-08 08:51:55.599513007 -0400 EDT m=+0.086835856 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35056]: This container is intended for podman CI testing\nJul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:55 managed-node1 podman[35060]: 2025-07-08 08:51:55.635383437 -0400 EDT m=+0.022862791 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:55 managed-node1 podman[35060]: 2025-07-08 08:51:55.647910671 -0400 EDT m=+0.035389965 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4045.\nJul 08 08:51:55 managed-node1 podman[35060]: 2025-07-08 08:51:55.691636435 -0400 EDT m=+0.079115737 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:55 managed-node1 podman[35060]: 2025-07-08 08:51:55.695110225 -0400 EDT m=+0.082589576 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:55 managed-node1 conmon[35071]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35071]: This container is intended for podman CI testing\nJul 08 08:51:55 managed-node1 podman[35075]: 2025-07-08 08:51:55.728779782 -0400 EDT m=+0.022198450 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:55 managed-node1 podman[35075]: 2025-07-08 08:51:55.741390318 -0400 EDT m=+0.034808950 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4050.\nJul 08 08:51:55 managed-node1 podman[35075]: 2025-07-08 08:51:55.789562384 -0400 EDT m=+0.082981033 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:55 managed-node1 podman[35075]: 2025-07-08 08:51:55.793529932 -0400 EDT m=+0.086948620 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35086]: This container is intended for podman CI testing\nJul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:55 managed-node1 podman[35090]: 2025-07-08 08:51:55.830769937 -0400 EDT m=+0.023577778 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:55 managed-node1 podman[35090]: 2025-07-08 08:51:55.843324486 -0400 EDT m=+0.036132283 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4055.\nJul 08 08:51:55 managed-node1 podman[35090]: 2025-07-08 08:51:55.894903684 -0400 EDT m=+0.087711555 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:55 managed-node1 podman[35090]: 2025-07-08 08:51:55.898507737 -0400 EDT m=+0.091315579 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35101]: This container is intended for podman CI testing\nJul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:55 managed-node1 conmon[35101]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:55 managed-node1 podman[35105]: 2025-07-08 08:51:55.930536412 -0400 EDT m=+0.021767162 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:55 managed-node1 podman[35105]: 2025-07-08 08:51:55.943827468 -0400 EDT m=+0.035058234 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:55 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4060.\nJul 08 08:51:55 managed-node1 podman[35105]: 2025-07-08 08:51:55.992851538 -0400 EDT m=+0.084082447 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:55 managed-node1 podman[35105]: 2025-07-08 08:51:55.996511805 -0400 EDT m=+0.087742643 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:55 managed-node1 auth_test_1_kube-auth_test_1_kube[35117]: This container is intended for podman CI testing\nJul 08 08:51:55 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:55 managed-node1 conmon[35117]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:56 managed-node1 podman[35121]: 2025-07-08 08:51:56.032338614 -0400 EDT m=+0.022974943 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:56 managed-node1 podman[35121]: 2025-07-08 08:51:56.044592593 -0400 EDT m=+0.035228838 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4065.\nJul 08 08:51:56 managed-node1 podman[35121]: 2025-07-08 08:51:56.092083577 -0400 EDT m=+0.082719908 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35133]: This container is intended for podman CI testing\nJul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:56 managed-node1 conmon[35133]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:56 managed-node1 podman[35121]: 2025-07-08 08:51:56.100143213 -0400 EDT m=+0.090779556 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:56 managed-node1 podman[35137]: 2025-07-08 08:51:56.1295111 -0400 EDT m=+0.023708974 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:56 managed-node1 podman[35137]: 2025-07-08 08:51:56.142683417 -0400 EDT m=+0.036881232 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4070.\nJul 08 08:51:56 managed-node1 podman[35137]: 2025-07-08 08:51:56.190097155 -0400 EDT m=+0.084295035 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35148]: This container is intended for podman CI testing\nJul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:56 managed-node1 podman[35137]: 2025-07-08 08:51:56.194245804 -0400 EDT m=+0.088443924 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:56 managed-node1 podman[35152]: 2025-07-08 08:51:56.226636924 -0400 EDT m=+0.023968011 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:56 managed-node1 podman[35152]: 2025-07-08 08:51:56.239370854 -0400 EDT m=+0.036701891 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4075.\nJul 08 08:51:56 managed-node1 podman[35152]: 2025-07-08 08:51:56.28775851 -0400 EDT m=+0.085089601 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:56 managed-node1 podman[35152]: 2025-07-08 08:51:56.290845272 -0400 EDT m=+0.088176435 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35163]: This container is intended for podman CI testing\nJul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:56 managed-node1 conmon[35163]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:56 managed-node1 podman[35167]: 2025-07-08 08:51:56.326445838 -0400 EDT m=+0.023013479 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:56 managed-node1 podman[35167]: 2025-07-08 08:51:56.33903617 -0400 EDT m=+0.035603761 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4080.\nJul 08 08:51:56 managed-node1 podman[35167]: 2025-07-08 08:51:56.392289413 -0400 EDT m=+0.088857050 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35178]: This container is intended for podman CI testing\nJul 08 08:51:56 managed-node1 podman[35167]: 2025-07-08 08:51:56.396104065 -0400 EDT m=+0.092671932 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:56 managed-node1 conmon[35178]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:56 managed-node1 podman[35182]: 2025-07-08 08:51:56.430712998 -0400 EDT m=+0.021670054 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:56 managed-node1 podman[35182]: 2025-07-08 08:51:56.443484263 -0400 EDT m=+0.034441280 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4085.\nJul 08 08:51:56 managed-node1 podman[35182]: 2025-07-08 08:51:56.492719275 -0400 EDT m=+0.083676318 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:56 managed-node1 podman[35182]: 2025-07-08 08:51:56.495965431 -0400 EDT m=+0.086922524 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35193]: This container is intended for podman CI testing\nJul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:56 managed-node1 conmon[35193]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:56 managed-node1 podman[35197]: 2025-07-08 08:51:56.536373017 -0400 EDT m=+0.027245791 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:56 managed-node1 podman[35197]: 2025-07-08 08:51:56.549889706 -0400 EDT m=+0.040762533 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4090.\nJul 08 08:51:56 managed-node1 podman[35197]: 2025-07-08 08:51:56.610056147 -0400 EDT m=+0.100928950 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35231]: This container is intended for podman CI testing\nJul 08 08:51:56 managed-node1 conmon[35231]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:56 managed-node1 podman[35197]: 2025-07-08 08:51:56.617514279 -0400 EDT m=+0.108386978 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:56 managed-node1 podman[35264]: 2025-07-08 08:51:56.669979715 -0400 EDT m=+0.043157775 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:56 managed-node1 podman[35264]: 2025-07-08 08:51:56.68324339 -0400 EDT m=+0.056421529 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4095.\nJul 08 08:51:56 managed-node1 podman[35264]: 2025-07-08 08:51:56.741763648 -0400 EDT m=+0.114941702 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:56 managed-node1 podman[35264]: 2025-07-08 08:51:56.745717027 -0400 EDT m=+0.118895041 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35322]: This container is intended for podman CI testing\nJul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:56 managed-node1 conmon[35322]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:56 managed-node1 podman[35351]: 2025-07-08 08:51:56.804720401 -0400 EDT m=+0.043304467 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:56 managed-node1 podman[35351]: 2025-07-08 08:51:56.818086586 -0400 EDT m=+0.056670229 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:56 managed-node1 python3.9[35352]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:51:56 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4100.\nJul 08 08:51:56 managed-node1 podman[35351]: 2025-07-08 08:51:56.921640357 -0400 EDT m=+0.160224188 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:56 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:56 managed-node1 auth_test_1_kube-auth_test_1_kube[35364]: This container is intended for podman CI testing\nJul 08 08:51:56 managed-node1 conmon[35364]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:56 managed-node1 podman[35351]: 2025-07-08 08:51:56.929562868 -0400 EDT m=+0.168146477 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:56 managed-node1 podman[35370]: 2025-07-08 08:51:56.978581974 -0400 EDT m=+0.040551987 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:56 managed-node1 podman[35370]: 2025-07-08 08:51:56.991489009 -0400 EDT m=+0.053458984 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4105.\nJul 08 08:51:57 managed-node1 podman[35370]: 2025-07-08 08:51:57.041682588 -0400 EDT m=+0.103652603 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35405]: This container is intended for podman CI testing\nJul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:57 managed-node1 podman[35370]: 2025-07-08 08:51:57.046108774 -0400 EDT m=+0.108078815 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:57 managed-node1 podman[35409]: 2025-07-08 08:51:57.080657767 -0400 EDT m=+0.025552325 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:57 managed-node1 podman[35409]: 2025-07-08 08:51:57.093888586 -0400 EDT m=+0.038783111 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4110.\nJul 08 08:51:57 managed-node1 podman[35409]: 2025-07-08 08:51:57.139013153 -0400 EDT m=+0.083907706 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35420]: This container is intended for podman CI testing\nJul 08 08:51:57 managed-node1 podman[35409]: 2025-07-08 08:51:57.144376284 -0400 EDT m=+0.089270897 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:57 managed-node1 podman[35424]: 2025-07-08 08:51:57.181570699 -0400 EDT m=+0.027520626 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:57 managed-node1 podman[35424]: 2025-07-08 08:51:57.194248987 -0400 EDT m=+0.040198878 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4115.\nJul 08 08:51:57 managed-node1 podman[35424]: 2025-07-08 08:51:57.242720898 -0400 EDT m=+0.088670798 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:57 managed-node1 podman[35424]: 2025-07-08 08:51:57.245785923 -0400 EDT m=+0.091735824 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35435]: This container is intended for podman CI testing\nJul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:57 managed-node1 podman[35439]: 2025-07-08 08:51:57.282621332 -0400 EDT m=+0.023409359 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:57 managed-node1 podman[35439]: 2025-07-08 08:51:57.295172051 -0400 EDT m=+0.035960073 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4120.\nJul 08 08:51:57 managed-node1 podman[35439]: 2025-07-08 08:51:57.34336529 -0400 EDT m=+0.084153346 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:57 managed-node1 podman[35439]: 2025-07-08 08:51:57.346573648 -0400 EDT m=+0.087361807 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35450]: This container is intended for podman CI testing\nJul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:57 managed-node1 podman[35454]: 2025-07-08 08:51:57.379900846 -0400 EDT m=+0.022316677 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:57 managed-node1 podman[35454]: 2025-07-08 08:51:57.392561635 -0400 EDT m=+0.034977428 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4125.\nJul 08 08:51:57 managed-node1 podman[35454]: 2025-07-08 08:51:57.450331168 -0400 EDT m=+0.092747058 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:57 managed-node1 podman[35454]: 2025-07-08 08:51:57.453552372 -0400 EDT m=+0.095968236 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35465]: This container is intended for podman CI testing\nJul 08 08:51:57 managed-node1 conmon[35465]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:57 managed-node1 podman[35469]: 2025-07-08 08:51:57.489276437 -0400 EDT m=+0.023295529 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:57 managed-node1 podman[35469]: 2025-07-08 08:51:57.501975537 -0400 EDT m=+0.035994590 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4130.\nJul 08 08:51:57 managed-node1 podman[35469]: 2025-07-08 08:51:57.550665142 -0400 EDT m=+0.084684211 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:57 managed-node1 podman[35469]: 2025-07-08 08:51:57.553825967 -0400 EDT m=+0.087845107 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35480]: This container is intended for podman CI testing\nJul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:57 managed-node1 podman[35484]: 2025-07-08 08:51:57.58960223 -0400 EDT m=+0.022372847 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:57 managed-node1 podman[35484]: 2025-07-08 08:51:57.605935724 -0400 EDT m=+0.038706383 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4135.\nJul 08 08:51:57 managed-node1 podman[35484]: 2025-07-08 08:51:57.665840337 -0400 EDT m=+0.098611026 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:57 managed-node1 podman[35484]: 2025-07-08 08:51:57.669206337 -0400 EDT m=+0.101977039 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35521]: This container is intended for podman CI testing\nJul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:57 managed-node1 podman[35551]: 2025-07-08 08:51:57.723078909 -0400 EDT m=+0.044059134 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:57 managed-node1 podman[35551]: 2025-07-08 08:51:57.738242838 -0400 EDT m=+0.059223139 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4140.\nJul 08 08:51:57 managed-node1 podman[35551]: 2025-07-08 08:51:57.795605487 -0400 EDT m=+0.116585731 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:57 managed-node1 podman[35551]: 2025-07-08 08:51:57.802164945 -0400 EDT m=+0.123145103 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35610]: This container is intended for podman CI testing\nJul 08 08:51:57 managed-node1 conmon[35610]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:57 managed-node1 podman[35640]: 2025-07-08 08:51:57.861994014 -0400 EDT m=+0.040045330 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:57 managed-node1 podman[35640]: 2025-07-08 08:51:57.878110989 -0400 EDT m=+0.056162178 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:57 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4145.\nJul 08 08:51:57 managed-node1 podman[35640]: 2025-07-08 08:51:57.934381076 -0400 EDT m=+0.112432215 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:57 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:57 managed-node1 auth_test_1_kube-auth_test_1_kube[35652]: This container is intended for podman CI testing\nJul 08 08:51:57 managed-node1 conmon[35652]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:57 managed-node1 podman[35640]: 2025-07-08 08:51:57.942413628 -0400 EDT m=+0.120464749 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:57 managed-node1 python3.9[35639]: ansible-ansible.legacy.command Invoked with _raw_params=systemd-escape --template podman-kube@.service /etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:51:57 managed-node1 podman[35656]: 2025-07-08 08:51:57.99609442 -0400 EDT m=+0.039690418 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:58 managed-node1 podman[35656]: 2025-07-08 08:51:58.015367809 -0400 EDT m=+0.058963962 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:58 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4150.\nJul 08 08:51:58 managed-node1 podman[35656]: 2025-07-08 08:51:58.071702173 -0400 EDT m=+0.115298222 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:58 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:58 managed-node1 auth_test_1_kube-auth_test_1_kube[35691]: This container is intended for podman CI testing\nJul 08 08:51:58 managed-node1 podman[35656]: 2025-07-08 08:51:58.081398246 -0400 EDT m=+0.124994462 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:58 managed-node1 podman[35696]: 2025-07-08 08:51:58.118603324 -0400 EDT m=+0.029031609 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:58 managed-node1 podman[35696]: 2025-07-08 08:51:58.132892828 -0400 EDT m=+0.043321060 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:58 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4155.\nJul 08 08:51:58 managed-node1 podman[35696]: 2025-07-08 08:51:58.179673011 -0400 EDT m=+0.090101285 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:58 managed-node1 podman[35696]: 2025-07-08 08:51:58.182883548 -0400 EDT m=+0.093311839 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:58 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:58 managed-node1 auth_test_1_kube-auth_test_1_kube[35707]: This container is intended for podman CI testing\nJul 08 08:51:58 managed-node1 conmon[35707]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:58 managed-node1 podman[35712]: 2025-07-08 08:51:58.228222867 -0400 EDT m=+0.032629915 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:58 managed-node1 podman[35712]: 2025-07-08 08:51:58.241845369 -0400 EDT m=+0.046252354 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:58 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4160.\nJul 08 08:51:58 managed-node1 podman[35712]: 2025-07-08 08:51:58.306194824 -0400 EDT m=+0.110602009 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:58 managed-node1 podman[35712]: 2025-07-08 08:51:58.311409402 -0400 EDT m=+0.115816394 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:58 managed-node1 auth_test_1_kube-auth_test_1_kube[35775]: This container is intended for podman CI testing\nJul 08 08:51:58 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:58 managed-node1 podman[35800]: 2025-07-08 08:51:58.350154057 -0400 EDT m=+0.026407392 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:58 managed-node1 podman[35800]: 2025-07-08 08:51:58.365884381 -0400 EDT m=+0.042137951 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:58 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4165.\nJul 08 08:51:58 managed-node1 podman[35800]: 2025-07-08 08:51:58.428581949 -0400 EDT m=+0.104835493 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:58 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:58 managed-node1 auth_test_1_kube-auth_test_1_kube[35844]: This container is intended for podman CI testing\nJul 08 08:51:58 managed-node1 conmon[35844]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:58 managed-node1 podman[35800]: 2025-07-08 08:51:58.436307291 -0400 EDT m=+0.112560702 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:58 managed-node1 podman[35867]: 2025-07-08 08:51:58.487344628 -0400 EDT m=+0.043335121 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:58 managed-node1 podman[35867]: 2025-07-08 08:51:58.500661958 -0400 EDT m=+0.056652344 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:58 managed-node1 python3.9[35865]: ansible-systemd Invoked with name=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service scope=system state=stopped enabled=False daemon_reload=False daemon_reexec=False no_block=False force=None masked=None\nJul 08 08:51:58 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4170.\nJul 08 08:51:58 managed-node1 podman[35867]: 2025-07-08 08:51:58.703250209 -0400 EDT m=+0.259240736 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:58 managed-node1 auth_test_1_kube-auth_test_1_kube[35878]: This container is intended for podman CI testing\nJul 08 08:51:58 managed-node1 podman[35867]: 2025-07-08 08:51:58.708624472 -0400 EDT m=+0.264614833 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:58 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:58 managed-node1 systemd[1]: Reloading.\nJul 08 08:51:58 managed-node1 podman[35884]: 2025-07-08 08:51:58.78180345 -0400 EDT m=+0.056786185 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:58 managed-node1 podman[35884]: 2025-07-08 08:51:58.796379401 -0400 EDT m=+0.071362044 container restart e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test)\nJul 08 08:51:58 managed-node1 systemd-rc-local-generator[35907]: /etc/rc.d/rc.local is not marked executable, skipping.\nJul 08 08:51:58 managed-node1 systemd[1]: Stopping A template for running K8s workloads via podman-kube-play...\n\u2591\u2591 Subject: A stop job for unit podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service has begun execution\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A stop job for unit podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service has begun execution.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4175.\nJul 08 08:51:59 managed-node1 systemd[1]: Started libcrun container.\n\u2591\u2591 Subject: A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4176.\nJul 08 08:51:59 managed-node1 podman[35884]: 2025-07-08 08:51:59.019304564 -0400 EDT m=+0.294287240 container init e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0)\nJul 08 08:51:59 managed-node1 podman[35884]: 2025-07-08 08:51:59.022879659 -0400 EDT m=+0.297862353 container start e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage)\nJul 08 08:51:59 managed-node1 systemd[1]: libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope has successfully entered the 'dead' state.\nJul 08 08:51:59 managed-node1 auth_test_1_kube-auth_test_1_kube[35932]: This container is intended for podman CI testing\nJul 08 08:51:59 managed-node1 conmon[35932]: conmon e6081ddd78f036da452c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9.scope/container/memory.events\nJul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.048700954 -0400 EDT m=+0.061178515 pod stop 8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c (image=, name=auth_test_1_kube)\nJul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.050394331 -0400 EDT m=+0.062872020 container died e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:59 managed-node1 systemd[1]: var-lib-containers-storage-overlay-765a43272b7d1ee027833f1b9208178861d0063df0fe51067dc546f0ffe0a36b-merged.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay-765a43272b7d1ee027833f1b9208178861d0063df0fe51067dc546f0ffe0a36b-merged.mount has successfully entered the 'dead' state.\nJul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.096509549 -0400 EDT m=+0.108986982 container cleanup e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry)\nJul 08 08:51:59 managed-node1 systemd[1]: libpod-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a.scope has successfully entered the 'dead' state.\nJul 08 08:51:59 managed-node1 conmon[20816]: conmon 83c6924d2fc59b99e53d : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice/libpod-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a.scope/container/memory.events\nJul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.117019225 -0400 EDT m=+0.129496774 container died 83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a (image=, name=8e1f57b75507-infra, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:59 managed-node1 systemd[1]: run-r453f02fb4ae24dce84d3e8de47463c2f.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit run-r453f02fb4ae24dce84d3e8de47463c2f.scope has successfully entered the 'dead' state.\nJul 08 08:51:59 managed-node1 kernel: podman1: port 1(veth1) entered disabled state\nJul 08 08:51:59 managed-node1 kernel: veth1 (unregistering): left allmulticast mode\nJul 08 08:51:59 managed-node1 kernel: veth1 (unregistering): left promiscuous mode\nJul 08 08:51:59 managed-node1 kernel: podman1: port 1(veth1) entered disabled state\nJul 08 08:51:59 managed-node1 NetworkManager[642]: [1751979119.1744] device (podman1): state change: activated -> unmanaged (reason 'unmanaged', managed-type: 'removed')\nJul 08 08:51:59 managed-node1 systemd[1]: Starting Network Manager Script Dispatcher Service...\n\u2591\u2591 Subject: A start job for unit NetworkManager-dispatcher.service has begun execution\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit NetworkManager-dispatcher.service has begun execution.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4182.\nJul 08 08:51:59 managed-node1 systemd[1]: Started Network Manager Script Dispatcher Service.\n\u2591\u2591 Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit NetworkManager-dispatcher.service has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4182.\nJul 08 08:51:59 managed-node1 systemd[1]: run-netns-netns\\x2d776ce46f\\x2dcdf2\\x2d9675\\x2daf23\\x2dbc7393892a77.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit run-netns-netns\\x2d776ce46f\\x2dcdf2\\x2d9675\\x2daf23\\x2dbc7393892a77.mount has successfully entered the 'dead' state.\nJul 08 08:51:59 managed-node1 systemd[1]: var-lib-containers-storage-overlay\\x2dcontainers-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a-rootfs-merge.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay\\x2dcontainers-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a-rootfs-merge.mount has successfully entered the 'dead' state.\nJul 08 08:51:59 managed-node1 systemd[1]: var-lib-containers-storage-overlay\\x2dcontainers-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a-userdata-shm.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay\\x2dcontainers-83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a-userdata-shm.mount has successfully entered the 'dead' state.\nJul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.309923758 -0400 EDT m=+0.322401287 container cleanup 83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a (image=, name=8e1f57b75507-infra, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:59 managed-node1 systemd[1]: Removed slice cgroup machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice.\n\u2591\u2591 Subject: A stop job for unit machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice has finished\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A stop job for unit machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice has finished.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4248 and the job result is done.\nJul 08 08:51:59 managed-node1 systemd[1]: machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice: Consumed 1.763s CPU time.\n\u2591\u2591 Subject: Resources consumed by unit runtime\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice completed and consumed the indicated resources.\nJul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.33731008 -0400 EDT m=+0.349787550 container remove e6081ddd78f036da452c7b8407975d264cbc5a5eaa0c7abca268ae6502a0d8b9 (image=localhost:5000/libpod/testimage:20210610, name=auth_test_1_kube-auth_test_1_kube, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, created_by=test/system/build-testimage, io.buildah.version=1.21.0, io.containers.autoupdate=registry, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service, app=test, created_at=2021-06-10T18:55:36Z)\nJul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.365548153 -0400 EDT m=+0.378025627 container remove 83c6924d2fc59b99e53d51ccb007ba613d4d50762638fe2c73e9af426783fd1a (image=, name=8e1f57b75507-infra, pod_id=8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:59 managed-node1 systemd[1]: machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice: Failed to open /run/systemd/transient/machine-libpod_pod_8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c.slice: No such file or directory\nJul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.37391867 -0400 EDT m=+0.386396101 pod remove 8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c (image=, name=auth_test_1_kube)\nJul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.377980277 -0400 EDT m=+0.390457910 container kill 8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1 (image=, name=3ef6fcac6278-service, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:59 managed-node1 systemd[1]: libpod-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1.scope has successfully entered the 'dead' state.\nJul 08 08:51:59 managed-node1 conmon[20738]: conmon 8ab849217cf3dcfe0949 : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/libpod-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1.scope/container/memory.events\nJul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.384674549 -0400 EDT m=+0.397152102 container died 8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1 (image=, name=3ef6fcac6278-service, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:59 managed-node1 systemd[1]: var-lib-containers-storage-overlay\\x2dcontainers-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1-rootfs-merge.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay\\x2dcontainers-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1-rootfs-merge.mount has successfully entered the 'dead' state.\nJul 08 08:51:59 managed-node1 podman[35930]: 2025-07-08 08:51:59.442987705 -0400 EDT m=+0.455465175 container remove 8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1 (image=, name=3ef6fcac6278-service, PODMAN_SYSTEMD_UNIT=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service)\nJul 08 08:51:59 managed-node1 podman[35930]: Pods stopped:\nJul 08 08:51:59 managed-node1 podman[35930]: 8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c\nJul 08 08:51:59 managed-node1 podman[35930]: Pods removed:\nJul 08 08:51:59 managed-node1 podman[35930]: 8e1f57b7550726e0c526c801b73e11a271bba131bffe92a105d9eb4021952d2c\nJul 08 08:51:59 managed-node1 podman[35930]: Secrets removed:\nJul 08 08:51:59 managed-node1 podman[35930]: Volumes removed:\nJul 08 08:51:59 managed-node1 systemd[1]: podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service has successfully entered the 'dead' state.\nJul 08 08:51:59 managed-node1 systemd[1]: Stopped A template for running K8s workloads via podman-kube-play.\n\u2591\u2591 Subject: A stop job for unit podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service has finished\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A stop job for unit podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service has finished.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4175 and the job result is done.\nJul 08 08:51:59 managed-node1 systemd[1]: podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service: Consumed 37.552s CPU time.\n\u2591\u2591 Subject: Resources consumed by unit runtime\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service completed and consumed the indicated resources.\nJul 08 08:51:59 managed-node1 python3.9[36157]: ansible-stat Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:00 managed-node1 systemd[1]: var-lib-containers-storage-overlay\\x2dcontainers-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1-userdata-shm.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay\\x2dcontainers-8ab849217cf3dcfe094973e1524c2d438ea5a729d28bb8af71cdcffac2d431f1-userdata-shm.mount has successfully entered the 'dead' state.\nJul 08 08:52:00 managed-node1 python3.9[36308]: ansible-containers.podman.podman_play Invoked with state=absent kube_file=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml executable=podman annotation=None kube_file_content=None authfile=None build=None cert_dir=None configmap=None context_dir=None seccomp_profile_root=None username=None password=NOT_LOGGING_PARAMETER log_driver=None log_opt=None network=None tls_verify=None debug=None quiet=None recreate=None userns=None log_level=None quadlet_dir=None quadlet_filename=None quadlet_file_mode=None quadlet_options=None\nJul 08 08:52:00 managed-node1 python3.9[36308]: ansible-containers.podman.podman_play version: 5.5.1, kube file /etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml\nJul 08 08:52:00 managed-node1 python3.9[36471]: ansible-file Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:02 managed-node1 python3.9[36620]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:03 managed-node1 python3.9[36771]: ansible-systemd Invoked with name=auth_test_1_quadlet.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None\nJul 08 08:52:03 managed-node1 systemd[1]: Reloading.\nJul 08 08:52:03 managed-node1 systemd-rc-local-generator[36791]: /etc/rc.d/rc.local is not marked executable, skipping.\nJul 08 08:52:03 managed-node1 python3.9[36956]: ansible-stat Invoked with path=/etc/containers/systemd/auth_test_1_quadlet.container follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:04 managed-node1 python3.9[37256]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:07 managed-node1 python3.9[37555]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:08 managed-node1 python3.9[37710]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:09 managed-node1 systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state.\nJul 08 08:52:10 managed-node1 python3.9[37861]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:12 managed-node1 python3.9[38012]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:14 managed-node1 python3.9[38163]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:15 managed-node1 python3.9[38314]: ansible-ansible.legacy.command Invoked with _raw_params=systemd-escape --template podman-kube@.service /etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:16 managed-node1 python3.9[38464]: ansible-systemd Invoked with name=podman-kube@-etc-containers-ansible\\x2dkubernetes.d-auth_test_1_kube.yml.service scope=system state=stopped enabled=False daemon_reload=False daemon_reexec=False no_block=False force=None masked=None\nJul 08 08:52:17 managed-node1 python3.9[38615]: ansible-stat Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:17 managed-node1 python3.9[38764]: ansible-file Invoked with path=/etc/containers/ansible-kubernetes.d/auth_test_1_kube.yml state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:19 managed-node1 python3.9[38913]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:20 managed-node1 python3.9[39064]: ansible-systemd Invoked with name=auth_test_1_quadlet.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None\nJul 08 08:52:20 managed-node1 systemd[1]: Reloading.\nJul 08 08:52:20 managed-node1 systemd-rc-local-generator[39087]: /etc/rc.d/rc.local is not marked executable, skipping.\nJul 08 08:52:21 managed-node1 python3.9[39249]: ansible-stat Invoked with path=/etc/containers/systemd/auth_test_1_quadlet.container follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:22 managed-node1 python3.9[39549]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:24 managed-node1 python3.9[39848]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:25 managed-node1 python3.9[40003]: ansible-getent Invoked with database=passwd key=auth_test_user1 fail_key=False service=None split=None\nJul 08 08:52:25 managed-node1 python3.9[40153]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:26 managed-node1 python3.9[40303]: ansible-user Invoked with name=auth_test_user1 state=absent non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on managed-node1 update_password=always uid=None group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None password_expire_warn=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None\nJul 08 08:52:26 managed-node1 python3.9[40453]: ansible-file Invoked with path=/home/auth_test_user1 state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:27 managed-node1 python3.9[40602]: ansible-ansible.legacy.command Invoked with _raw_params=podman inspect podman_registry --format '{{range .}}{{range .Mounts}}{{if eq .Type \"volume\"}}{{.Name}}{{end}}{{end}}{{end}}' _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:27 managed-node1 python3.9[40759]: ansible-ansible.legacy.command Invoked with _raw_params=podman rm -f podman_registry _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:27 managed-node1 systemd[1]: libpod-2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354.scope has successfully entered the 'dead' state.\nJul 08 08:52:27 managed-node1 podman[40760]: 2025-07-08 08:52:27.930931198 -0400 EDT m=+0.044482479 container died 2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354 (image=quay.io/libpod/registry:2.8.2, name=podman_registry)\nJul 08 08:52:27 managed-node1 kernel: podman0: port 1(veth0) entered disabled state\nJul 08 08:52:27 managed-node1 kernel: veth0 (unregistering): left allmulticast mode\nJul 08 08:52:27 managed-node1 kernel: veth0 (unregistering): left promiscuous mode\nJul 08 08:52:27 managed-node1 kernel: podman0: port 1(veth0) entered disabled state\nJul 08 08:52:27 managed-node1 NetworkManager[642]: [1751979147.9779] device (podman0): state change: activated -> unmanaged (reason 'unmanaged', managed-type: 'removed')\nJul 08 08:52:27 managed-node1 systemd[1]: Starting Network Manager Script Dispatcher Service...\n\u2591\u2591 Subject: A start job for unit NetworkManager-dispatcher.service has begun execution\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit NetworkManager-dispatcher.service has begun execution.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4250.\nJul 08 08:52:27 managed-node1 systemd[1]: Started Network Manager Script Dispatcher Service.\n\u2591\u2591 Subject: A start job for unit NetworkManager-dispatcher.service has finished successfully\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 A start job for unit NetworkManager-dispatcher.service has finished successfully.\n\u2591\u2591 \n\u2591\u2591 The job identifier is 4250.\nJul 08 08:52:28 managed-node1 systemd[1]: run-netns-netns\\x2dcbdb4e71\\x2d6d48\\x2da1b0\\x2d498d\\x2d53f813dda638.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit run-netns-netns\\x2dcbdb4e71\\x2d6d48\\x2da1b0\\x2d498d\\x2d53f813dda638.mount has successfully entered the 'dead' state.\nJul 08 08:52:28 managed-node1 systemd[1]: var-lib-containers-storage-overlay\\x2dcontainers-2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354-userdata-shm.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay\\x2dcontainers-2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354-userdata-shm.mount has successfully entered the 'dead' state.\nJul 08 08:52:28 managed-node1 systemd[1]: var-lib-containers-storage-overlay-7ad31fb747dc13efb3b22f585d1a06b75296c677706679c325a915f2d85e1e11-merged.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay-7ad31fb747dc13efb3b22f585d1a06b75296c677706679c325a915f2d85e1e11-merged.mount has successfully entered the 'dead' state.\nJul 08 08:52:28 managed-node1 podman[40760]: 2025-07-08 08:52:28.143410392 -0400 EDT m=+0.256961352 container remove 2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354 (image=quay.io/libpod/registry:2.8.2, name=podman_registry)\nJul 08 08:52:28 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.\nJul 08 08:52:28 managed-node1 systemd[1]: libpod-conmon-2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354.scope: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit libpod-conmon-2ad8371c24bb2603ad38801c033c94b9b279614166f08bda589973bf89f28354.scope has successfully entered the 'dead' state.\nJul 08 08:52:28 managed-node1 python3.9[40973]: ansible-ansible.legacy.command Invoked with _raw_params=podman volume rm c5dabb915f4c30f58eb29c345d18cc09b42d61980a614c3a9fecfd122d7d7dbb _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:28 managed-node1 podman[40974]: 2025-07-08 08:52:28.605123907 -0400 EDT m=+0.025267136 volume remove c5dabb915f4c30f58eb29c345d18cc09b42d61980a614c3a9fecfd122d7d7dbb\nJul 08 08:52:28 managed-node1 python3.9[41130]: ansible-file Invoked with path=/tmp/lsr_dbjlhh8o_podman state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:29 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.\nJul 08 08:52:31 managed-node1 python3.9[41328]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d\nJul 08 08:52:32 managed-node1 python3.9[41503]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:32 managed-node1 python3.9[41652]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:34 managed-node1 python3.9[41950]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:35 managed-node1 python3.9[42105]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None\nJul 08 08:52:36 managed-node1 python3.9[42255]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:38 managed-node1 systemd[1]: NetworkManager-dispatcher.service: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state.\nJul 08 08:52:38 managed-node1 python3.9[42406]: ansible-tempfile Invoked with state=directory prefix=lsr_podman_config_ suffix= path=None\nJul 08 08:52:38 managed-node1 python3.9[42555]: ansible-ansible.legacy.command Invoked with _raw_params=tar --ignore-failed-read -c -P -v -p -f /tmp/lsr_podman_config_6ihb6fh6/backup.tar /etc/containers/containers.conf.d/50-systemroles.conf /etc/containers/registries.conf.d/50-systemroles.conf /etc/containers/storage.conf /etc/containers/policy.json _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:39 managed-node1 python3.9[42705]: ansible-user Invoked with name=user1 state=present non_unique=False force=False remove=False create_home=True system=False move_home=False append=False ssh_key_bits=0 ssh_key_type=rsa ssh_key_comment=ansible-generated on managed-node1 update_password=always uid=None group=None groups=None comment=None home=None shell=None password=NOT_LOGGING_PARAMETER login_class=None password_expire_max=None password_expire_min=None password_expire_warn=None hidden=None seuser=None skeleton=None generate_ssh_key=None ssh_key_file=None ssh_key_passphrase=NOT_LOGGING_PARAMETER expires=None password_lock=None local=None profile=None authorization=None role=None umask=None\nJul 08 08:52:39 managed-node1 useradd[42707]: new group: name=user1, GID=1000\nJul 08 08:52:39 managed-node1 useradd[42707]: new user: name=user1, UID=1000, GID=1000, home=/home/user1, shell=/bin/bash, from=/dev/pts/0\nJul 08 08:52:41 managed-node1 python3.9[43011]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:42 managed-node1 python3.9[43167]: ansible-getent Invoked with database=passwd key=user1 fail_key=False service=None split=None\nJul 08 08:52:42 managed-node1 python3.9[43317]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:42 managed-node1 python3.9[43468]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:43 managed-node1 python3.9[43618]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:44 managed-node1 python3.9[43768]: ansible-file Invoked with path=/home/user1/.config/containers/containers.conf.d state=directory owner=user1 group=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:44 managed-node1 python3.9[43917]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:52:45 managed-node1 python3.9[44037]: ansible-ansible.legacy.copy Invoked with dest=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf owner=user1 group=user1 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979164.4055185-13192-99143701796826/.source.conf _original_basename=.fz16347w follow=False checksum=b1776092f2908d76e11fd6af87267469b2c17d5a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:45 managed-node1 python3.9[44186]: ansible-file Invoked with path=/home/user1/.config/containers/registries.conf.d state=directory owner=user1 group=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:46 managed-node1 python3.9[44335]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:52:46 managed-node1 python3.9[44455]: ansible-ansible.legacy.copy Invoked with dest=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf owner=user1 group=user1 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979165.7483666-13235-139178079209891/.source.conf _original_basename=.a1xet5pm follow=False checksum=fde25488ce7040f1639af7bfc88ed125318cc0b0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:46 managed-node1 python3.9[44604]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 group=user1 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:47 managed-node1 python3.9[44753]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:52:47 managed-node1 python3.9[44873]: ansible-ansible.legacy.copy Invoked with dest=/home/user1/.config/containers/storage.conf owner=user1 group=user1 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979166.9236264-13266-54800067456402/.source.conf _original_basename=.7t8_db15 follow=False checksum=38f015f4780579bd388dd955b42916199fd7fe19 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:47 managed-node1 python3.9[45022]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 group=user1 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:48 managed-node1 python3.9[45171]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:48 managed-node1 python3.9[45320]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:52:49 managed-node1 python3.9[45440]: ansible-ansible.legacy.copy Invoked with dest=/home/user1/.config/containers/policy.json owner=user1 group=user1 mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979168.451584-13317-138138508589651/.source.json _original_basename=.tq0hx2zm follow=False checksum=6746c079ad563b735fc39f73d4876654b80b0a0d backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:49 managed-node1 python3.9[45589]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:50 managed-node1 python3.9[45740]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:50 managed-node1 python3.9[45890]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:51 managed-node1 python3.9[46040]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 group=user1 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:52:52 managed-node1 python3.9[46458]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:53 managed-node1 python3.9[46609]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:53 managed-node1 python3.9[46759]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:54 managed-node1 python3.9[46909]: ansible-stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:54 managed-node1 python3.9[47060]: ansible-stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:55 managed-node1 python3.9[47211]: ansible-stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:55 managed-node1 python3.9[47362]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:55 managed-node1 python3.9[47513]: ansible-ansible.legacy.command Invoked with _raw_params=grep 'container_name_as_hostname[ ]*=[ ]*true' /home/user1/.config/containers/containers.conf.d/50-systemroles.conf _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:57 managed-node1 python3.9[47812]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:58 managed-node1 python3.9[47967]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:52:59 managed-node1 python3.9[48118]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:52:59 managed-node1 python3.9[48268]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:53:00 managed-node1 python3.9[48418]: ansible-file Invoked with path=/home/user1/.config/containers/containers.conf.d state=directory owner=user1 group=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:00 managed-node1 python3.9[48567]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:01 managed-node1 python3.9[48642]: ansible-ansible.legacy.file Invoked with owner=user1 group=user1 mode=0644 dest=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf _original_basename=.os5cs4_x recurse=False state=file path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:01 managed-node1 python3.9[48791]: ansible-file Invoked with path=/home/user1/.config/containers/registries.conf.d state=directory owner=user1 group=user1 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:02 managed-node1 python3.9[48940]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:02 managed-node1 python3.9[49015]: ansible-ansible.legacy.file Invoked with owner=user1 group=user1 mode=0644 dest=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf _original_basename=.9twz1wls recurse=False state=file path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:03 managed-node1 python3.9[49164]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 group=user1 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:03 managed-node1 python3.9[49313]: ansible-ansible.legacy.stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:03 managed-node1 python3.9[49388]: ansible-ansible.legacy.file Invoked with owner=user1 group=user1 mode=0644 dest=/home/user1/.config/containers/storage.conf _original_basename=.u72jitw8 recurse=False state=file path=/home/user1/.config/containers/storage.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:04 managed-node1 python3.9[49537]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 group=user1 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:04 managed-node1 python3.9[49686]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:04 managed-node1 python3.9[49837]: ansible-slurp Invoked with path=/home/user1/.config/containers/policy.json src=/home/user1/.config/containers/policy.json\nJul 08 08:53:05 managed-node1 python3.9[49986]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:06 managed-node1 python3.9[50137]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:53:06 managed-node1 python3.9[50287]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:53:07 managed-node1 python3.9[50437]: ansible-file Invoked with path=/home/user1/.config/containers state=directory owner=user1 group=user1 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:09 managed-node1 python3.9[50810]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:09 managed-node1 python3.9[50961]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:53:10 managed-node1 python3.9[51111]: ansible-ansible.legacy.command Invoked with _raw_params=getsubids -g user1 _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:53:10 managed-node1 python3.9[51261]: ansible-stat Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:11 managed-node1 python3.9[51412]: ansible-stat Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:11 managed-node1 python3.9[51563]: ansible-stat Invoked with path=/home/user1/.config/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:12 managed-node1 python3.9[51714]: ansible-stat Invoked with path=/home/user1/.config/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:13 managed-node1 python3.9[52014]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:53:14 managed-node1 python3.9[52169]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None\nJul 08 08:53:15 managed-node1 python3.9[52319]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:15 managed-node1 python3.9[52470]: ansible-file Invoked with path=/etc/containers/containers.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:16 managed-node1 python3.9[52619]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:16 managed-node1 python3.9[52739]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/containers.conf.d/50-systemroles.conf owner=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979196.068439-14128-196243318775352/.source.conf _original_basename=.9z01ulnk follow=False checksum=b1776092f2908d76e11fd6af87267469b2c17d5a backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:17 managed-node1 python3.9[52888]: ansible-file Invoked with path=/etc/containers/registries.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:17 managed-node1 python3.9[53037]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:17 managed-node1 python3.9[53157]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/registries.conf.d/50-systemroles.conf owner=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979197.2308524-14157-252335483909182/.source.conf _original_basename=.wkt7vbqw follow=False checksum=fde25488ce7040f1639af7bfc88ed125318cc0b0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:18 managed-node1 python3.9[53306]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:18 managed-node1 python3.9[53455]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:18 managed-node1 python3.9[53577]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/storage.conf owner=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979198.3939052-14186-108211027079466/.source.conf _original_basename=.6_l0ebnm follow=False checksum=38f015f4780579bd388dd955b42916199fd7fe19 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:19 managed-node1 python3.9[53726]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:19 managed-node1 python3.9[53875]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:20 managed-node1 python3.9[54026]: ansible-slurp Invoked with path=/etc/containers/policy.json src=/etc/containers/policy.json\nJul 08 08:53:20 managed-node1 python3.9[54175]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:21 managed-node1 python3.9[54297]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/policy.json owner=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979200.4464686-14254-119110580364058/.source.json _original_basename=.9grd0e5j follow=False checksum=6746c079ad563b735fc39f73d4876654b80b0a0d backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:21 managed-node1 python3.9[54446]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:22 managed-node1 python3.9[54597]: ansible-file Invoked with path=/root/.config/containers state=directory owner=root group=0 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:24 managed-node1 python3.9[55017]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:25 managed-node1 python3.9[55168]: ansible-stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:26 managed-node1 python3.9[55319]: ansible-stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:26 managed-node1 python3.9[55470]: ansible-stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:26 managed-node1 python3.9[55621]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:28 managed-node1 python3.9[55921]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:53:29 managed-node1 python3.9[56076]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:30 managed-node1 python3.9[56227]: ansible-file Invoked with path=/etc/containers/containers.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:30 managed-node1 python3.9[56376]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:31 managed-node1 python3.9[56451]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/containers.conf.d/50-systemroles.conf _original_basename=.lk6r9zg4 recurse=False state=file path=/etc/containers/containers.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:31 managed-node1 python3.9[56600]: ansible-file Invoked with path=/etc/containers/registries.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:32 managed-node1 python3.9[56749]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:32 managed-node1 python3.9[56824]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/registries.conf.d/50-systemroles.conf _original_basename=.0hqjoqag recurse=False state=file path=/etc/containers/registries.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:32 managed-node1 python3.9[56973]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:33 managed-node1 python3.9[57122]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:33 managed-node1 python3.9[57197]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/storage.conf _original_basename=.1ivx98ms recurse=False state=file path=/etc/containers/storage.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:33 managed-node1 python3.9[57346]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:34 managed-node1 python3.9[57495]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:34 managed-node1 python3.9[57646]: ansible-slurp Invoked with path=/etc/containers/policy.json src=/etc/containers/policy.json\nJul 08 08:53:35 managed-node1 python3.9[57795]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:36 managed-node1 python3.9[57946]: ansible-file Invoked with path=/root/.config/containers state=directory owner=root group=0 mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:37 managed-node1 python3.9[58319]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:38 managed-node1 python3.9[58470]: ansible-stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:39 managed-node1 python3.9[58621]: ansible-stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:39 managed-node1 python3.9[58772]: ansible-stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:40 managed-node1 python3.9[58923]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:40 managed-node1 python3.9[59074]: ansible-slurp Invoked with path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf src=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf\nJul 08 08:53:41 managed-node1 python3.9[59223]: ansible-slurp Invoked with path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf src=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf\nJul 08 08:53:41 managed-node1 python3.9[59372]: ansible-slurp Invoked with path=/home/user1/.config/containers/storage.conf src=/home/user1/.config/containers/storage.conf\nJul 08 08:53:42 managed-node1 python3.9[59521]: ansible-slurp Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf src=/etc/containers/containers.conf.d/50-systemroles.conf\nJul 08 08:53:42 managed-node1 python3.9[59670]: ansible-slurp Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf src=/etc/containers/registries.conf.d/50-systemroles.conf\nJul 08 08:53:43 managed-node1 python3.9[59819]: ansible-slurp Invoked with path=/etc/containers/storage.conf src=/etc/containers/storage.conf\nJul 08 08:53:45 managed-node1 python3.9[60117]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:53:46 managed-node1 python3.9[60273]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:47 managed-node1 python3.9[60424]: ansible-file Invoked with path=/etc/containers/containers.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:48 managed-node1 python3.9[60573]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:48 managed-node1 python3.9[60695]: ansible-ansible.legacy.copy Invoked with dest=/etc/containers/containers.conf.d/50-systemroles.conf owner=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1751979227.7911222-15133-177756245042776/.source.conf _original_basename=.eiu_kyxh follow=False checksum=9694c1d1c700a6435eecf4066b052584f4ee94c0 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:48 managed-node1 python3.9[60844]: ansible-file Invoked with path=/etc/containers/registries.conf.d state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:49 managed-node1 python3.9[60993]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/registries.conf.d/50-systemroles.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:49 managed-node1 python3.9[61068]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/registries.conf.d/50-systemroles.conf _original_basename=.gr7jiie7 recurse=False state=file path=/etc/containers/registries.conf.d/50-systemroles.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:50 managed-node1 python3.9[61217]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:50 managed-node1 python3.9[61366]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/storage.conf follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:53:51 managed-node1 python3.9[61441]: ansible-ansible.legacy.file Invoked with owner=root mode=0644 dest=/etc/containers/storage.conf _original_basename=.pcxrre0c recurse=False state=file path=/etc/containers/storage.conf force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:51 managed-node1 python3.9[61590]: ansible-file Invoked with path=/etc/containers state=directory owner=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:52 managed-node1 python3.9[61739]: ansible-stat Invoked with path=/etc/containers/policy.json follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:53:52 managed-node1 python3.9[61890]: ansible-slurp Invoked with path=/etc/containers/policy.json src=/etc/containers/policy.json\nJul 08 08:53:53 managed-node1 python3.9[62039]: ansible-slurp Invoked with path=/etc/containers/containers.conf.d/50-systemroles.conf src=/etc/containers/containers.conf.d/50-systemroles.conf\nJul 08 08:53:53 managed-node1 python3.9[62188]: ansible-file Invoked with state=absent path=/etc/containers/containers.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:54 managed-node1 python3.9[62337]: ansible-file Invoked with state=absent path=/etc/containers/registries.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:54 managed-node1 python3.9[62486]: ansible-file Invoked with state=absent path=/etc/containers/storage.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:54 managed-node1 python3.9[62635]: ansible-file Invoked with state=absent path=/etc/containers/policy.json recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:55 managed-node1 python3.9[62784]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/containers.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:55 managed-node1 python3.9[62933]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/registries.conf.d/50-systemroles.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:56 managed-node1 python3.9[63082]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/storage.conf recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:56 managed-node1 python3.9[63231]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/policy.json recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:56 managed-node1 python3.9[63380]: ansible-file Invoked with state=absent path=/root/.config/containers/auth.json recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:57 managed-node1 python3.9[63529]: ansible-file Invoked with state=absent path=/home/user1/.config/containers/auth.json recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:53:57 managed-node1 python3.9[63678]: ansible-ansible.legacy.command Invoked with _raw_params=tar xfvpP /tmp/lsr_podman_config_6ihb6fh6/backup.tar _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:53:57 managed-node1 python3.9[63828]: ansible-file Invoked with state=absent path=/tmp/lsr_podman_config_6ihb6fh6 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:54:00 managed-node1 python3.9[64026]: ansible-setup Invoked with gather_subset=['!all', '!min', 'distribution', 'distribution_major_version', 'distribution_version', 'os_family'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d\nJul 08 08:54:00 managed-node1 python3.9[64177]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:01 managed-node1 python3.9[64326]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:03 managed-node1 python3.9[64624]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:54:03 managed-node1 python3.9[64779]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None\nJul 08 08:54:04 managed-node1 python3.9[64929]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:07 managed-node1 python3.9[65129]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d\nJul 08 08:54:10 managed-node1 python3.9[65304]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:11 managed-node1 python3.9[65453]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:13 managed-node1 python3.9[65751]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:54:14 managed-node1 python3.9[65907]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None\nJul 08 08:54:14 managed-node1 python3.9[66057]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:19 managed-node1 python3.9[66257]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d\nJul 08 08:54:20 managed-node1 python3.9[66432]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:21 managed-node1 python3.9[66581]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:23 managed-node1 python3.9[66879]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:54:23 managed-node1 python3.9[67034]: ansible-getent Invoked with database=passwd key=root fail_key=False service=None split=None\nJul 08 08:54:24 managed-node1 python3.9[67184]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:27 managed-node1 python3.9[67335]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:28 managed-node1 python3.9[67486]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:54:29 managed-node1 python3.9[67635]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/nopull.container follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:54:29 managed-node1 python3.9[67755]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1751979269.1189396-16954-11122430624534/.source.container dest=/etc/containers/systemd/nopull.container owner=root group=0 mode=0644 follow=False _original_basename=systemd.j2 checksum=670d64fc68a9768edb20cad26df2acc703542d85 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:54:32 managed-node1 python3.9[68053]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:54:33 managed-node1 python3.9[68208]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:35 managed-node1 python3.9[68359]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:37 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.\nJul 08 08:54:37 managed-node1 podman[68519]: 2025-07-08 08:54:37.140415246 -0400 EDT m=+0.021147061 image pull-error this_is_a_bogus_image:latest short-name resolution enforced but cannot prompt without a TTY\nJul 08 08:54:37 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.\nJul 08 08:54:37 managed-node1 python3.9[68674]: ansible-file Invoked with path=/etc/containers/systemd state=directory owner=root group=0 mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:54:38 managed-node1 python3.9[68823]: ansible-ansible.legacy.stat Invoked with path=/etc/containers/systemd/bogus.container follow=False get_checksum=True get_size=False checksum_algorithm=sha1 get_mime=True get_attributes=True\nJul 08 08:54:38 managed-node1 python3.9[68943]: ansible-ansible.legacy.copy Invoked with src=/root/.ansible/tmp/ansible-tmp-1751979277.771543-17272-2966124504745/.source.container dest=/etc/containers/systemd/bogus.container owner=root group=0 mode=0644 follow=False _original_basename=systemd.j2 checksum=1d087e679d135214e8ac9ccaf33b2222916efb7f backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None\nJul 08 08:54:41 managed-node1 python3.9[69241]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:54:42 managed-node1 python3.9[69396]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:45 managed-node1 python3.9[69547]: ansible-stat Invoked with path=/usr/bin/getsubids follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:46 managed-node1 python3.9[69698]: ansible-systemd Invoked with name=nopull.service scope=system state=stopped enabled=False force=True daemon_reload=False daemon_reexec=False no_block=False masked=None\nJul 08 08:54:47 managed-node1 python3.9[69848]: ansible-stat Invoked with path=/etc/containers/systemd/nopull.container follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1\nJul 08 08:54:48 managed-node1 python3.9[70148]: ansible-ansible.legacy.command Invoked with _raw_params=set -x\n set -o pipefail\n exec 1>&2\n #podman volume rm --all\n #podman network prune -f\n podman volume ls\n podman network ls\n podman secret ls\n podman container ls\n podman pod ls\n podman images\n systemctl list-units | grep quadlet\n _uses_shell=True expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:54:48 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.\nJul 08 08:54:48 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.\nJul 08 08:54:49 managed-node1 systemd[1]: var-lib-containers-storage-overlay.mount: Deactivated successfully.\n\u2591\u2591 Subject: Unit succeeded\n\u2591\u2591 Defined-By: systemd\n\u2591\u2591 Support: https://access.redhat.com/support\n\u2591\u2591 \n\u2591\u2591 The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state.\nJul 08 08:54:50 managed-node1 python3.9[70491]: ansible-ansible.legacy.command Invoked with _raw_params=podman --version _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None\nJul 08 08:54:51 managed-node1 python3.9[70646]: ansible-getent Invoked with database=passwd key=user_quadlet_basic fail_key=False service=None split=None\nJul 08 08:54:51 managed-node1 python3.9[70796]: ansible-ansible.legacy.command Invoked with _raw_params=journalctl -ex _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None", "task_name": "Dump journal", "task_path": "/tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:336" } ] SYSTEM ROLES ERRORS END v1 TASKS RECAP ******************************************************************** Tuesday 08 July 2025 08:54:51 -0400 (0:00:00.443) 0:00:32.702 ********** =============================================================================== Gathering Facts --------------------------------------------------------- 1.16s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:9 fedora.linux_system_roles.podman : Gather the package facts ------------- 1.04s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.podman : Gather the package facts ------------- 0.92s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.podman : Ensure quadlet file is present ------- 0.86s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:75 fedora.linux_system_roles.podman : Gather the package facts ------------- 0.82s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.podman : Gather the package facts ------------- 0.81s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:6 fedora.linux_system_roles.podman : Stop and disable service ------------- 0.80s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/cleanup_quadlet_spec.yml:12 fedora.linux_system_roles.podman : Ensure quadlet file is present ------- 0.74s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:75 fedora.linux_system_roles.podman : Ensure container images are present --- 0.70s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:18 fedora.linux_system_roles.podman : Ensure the quadlet directory is present --- 0.56s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/create_update_quadlet_spec.yml:39 Debug3 ------------------------------------------------------------------ 0.55s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:270 fedora.linux_system_roles.podman : Get user information ----------------- 0.51s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:2 fedora.linux_system_roles.podman : Get podman version ------------------- 0.49s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 fedora.linux_system_roles.podman : Slurp quadlet file ------------------- 0.46s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/parse_quadlet_file.yml:6 fedora.linux_system_roles.podman : Get podman version ------------------- 0.45s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 fedora.linux_system_roles.podman : Check if system is ostree ------------ 0.45s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/set_vars.yml:11 Dump journal ------------------------------------------------------------ 0.44s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/tests/podman/tests_quadlet_basic.yml:336 fedora.linux_system_roles.podman : Get podman version ------------------- 0.43s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/main.yml:46 fedora.linux_system_roles.podman : See if getsubids exists -------------- 0.43s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31 fedora.linux_system_roles.podman : See if getsubids exists -------------- 0.42s /tmp/collections-4zP/ansible_collections/fedora/linux_system_roles/roles/podman/tasks/handle_user_group.yml:31