ansible-playbook 2.9.27 config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.9/site-packages/ansible executable location = /usr/local/bin/ansible-playbook python version = 3.9.19 (main, May 16 2024, 11:40:09) [GCC 8.5.0 20210514 (Red Hat 8.5.0-22)] No config file found; using defaults [WARNING]: running playbook inside collection fedora.linux_system_roles Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'jsonl', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_ntp_provider1.yml ********************************************** 1 plays in /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tests_ntp_provider1.yml PLAY [Configure NTP with default provider] ************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tests_ntp_provider1.yml:2 Monday 02 February 2026 18:19:03 -0500 (0:00:00.024) 0:00:00.024 ******* ok: [managed-node3] TASK [Common test setup tasks] ************************************************* task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tests_ntp_provider1.yml:11 Monday 02 February 2026 18:19:04 -0500 (0:00:00.872) 0:00:00.897 ******* included: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tasks/setup.yml for managed-node3 TASK [Run the role only to get vars needed for validation] ********************* task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tasks/setup.yml:3 Monday 02 February 2026 18:19:04 -0500 (0:00:00.017) 0:00:00.914 ******* TASK [fedora.linux_system_roles.timesync : Ensure ansible_facts used by role] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:2 Monday 02 February 2026 18:19:04 -0500 (0:00:00.036) 0:00:00.950 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Check if system is ostree] ********** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:10 Monday 02 February 2026 18:19:04 -0500 (0:00:00.051) 0:00:01.002 ******* ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.timesync : Set flag to indicate system is ostree] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:15 Monday 02 February 2026 18:19:05 -0500 (0:00:00.550) 0:00:01.552 ******* ok: [managed-node3] => { "ansible_facts": { "__timesync_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.timesync : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:22 Monday 02 February 2026 18:19:05 -0500 (0:00:00.042) 0:00:01.594 ******* ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.timesync : Set flag if transactional-update exists] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:27 Monday 02 February 2026 18:19:05 -0500 (0:00:00.374) 0:00:01.969 ******* ok: [managed-node3] => { "ansible_facts": { "__timesync_is_transactional": false }, "changed": false } TASK [fedora.linux_system_roles.timesync : Set platform/version specific variables] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:31 Monday 02 February 2026 18:19:05 -0500 (0:00:00.040) 0:00:02.010 ******* ok: [managed-node3] => { "ansible_facts": { "timesync_chrony_conf_path": "/etc/chrony.conf", "timesync_chrony_dhcp_sourcedir": "", "timesync_chrony_sysconfig_options": "", "timesync_chrony_sysconfig_path": "/etc/sysconfig/chronyd", "timesync_ntp_provider_os_default": "chrony", "timesync_ntp_sysconfig_path": "/etc/sysconfig/ntpd", "timesync_phc2sys_sysconfig_path": "/etc/sysconfig/phc2sys", "timesync_ptp4l_sysconfig_path": "/etc/sysconfig/ptp4l", "timesync_timemaster_config_path": "/etc/timemaster.conf" }, "ansible_included_var_files": [ "/tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/vars/CentOS_8.yml" ], "changed": false } TASK [fedora.linux_system_roles.timesync : Run systemctl] ********************** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:47 Monday 02 February 2026 18:19:05 -0500 (0:00:00.044) 0:00:02.054 ******* ok: [managed-node3] => { "changed": false, "cmd": [ "systemctl", "is-system-running" ], "delta": "0:00:00.006870", "end": "2026-02-02 18:19:06.019855", "failed_when_result": false, "rc": 0, "start": "2026-02-02 18:19:06.012985" } STDOUT: running TASK [fedora.linux_system_roles.timesync : Require installed systemd] ********** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:55 Monday 02 February 2026 18:19:06 -0500 (0:00:00.442) 0:00:02.496 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Set flag to indicate that systemd runtime operations are available] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:60 Monday 02 February 2026 18:19:06 -0500 (0:00:00.059) 0:00:02.556 ******* ok: [managed-node3] => { "ansible_facts": { "__timesync_is_booted": true }, "changed": false } META: end_host conditional evaluated to false, continuing execution for managed-node3 TASK [Ensure iproute for gathering default_ipv4 fact] ************************** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tasks/setup.yml:15 Monday 02 February 2026 18:19:06 -0500 (0:00:00.094) 0:00:02.651 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Ensure ansible_facts used by test] *************************************** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tasks/setup.yml:21 Monday 02 February 2026 18:19:20 -0500 (0:00:14.133) 0:00:16.784 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Debug] ******************************************************************* task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tasks/setup.yml:29 Monday 02 February 2026 18:19:20 -0500 (0:00:00.029) 0:00:16.814 ******* ok: [managed-node3] => {} MSG: facts { "_ansible_facts_gathered": true, "all_ipv4_addresses": [ "10.31.47.254" ], "all_ipv6_addresses": [ "fe80::6:8cff:fe5b:9bed" ], "ansible_local": {}, "apparmor": { "status": "disabled" }, "architecture": "x86_64", "bios_date": "08/24/2006", "bios_version": "4.11.amazon", "cmdline": { "BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-553.5.1.el8.x86_64", "crashkernel": "auto", "net.ifnames": "0", "quiet": true, "rhgb": true, "ro": true, "root": "UUID=fe591198-9082-4b15-9b62-e83518524cd2" }, "date_time": { "date": "2026-02-02", "day": "02", "epoch": "1770074344", "hour": "18", "iso8601": "2026-02-02T23:19:04Z", "iso8601_basic": "20260202T181904276290", "iso8601_basic_short": "20260202T181904", "iso8601_micro": "2026-02-02T23:19:04.276290Z", "minute": "19", "month": "02", "second": "04", "time": "18:19:04", "tz": "EST", "tz_offset": "-0500", "weekday": "Monday", "weekday_number": "1", "weeknumber": "05", "year": "2026" }, "default_ipv4": { "address": "10.31.47.254", "alias": "eth0", "broadcast": "10.31.47.255", "gateway": "10.31.44.1", "interface": "eth0", "macaddress": "02:06:8c:5b:9b:ed", "mtu": 9001, "netmask": "255.255.252.0", "network": "10.31.44.0", "type": "ether" }, "default_ipv6": {}, "device_links": { "ids": {}, "labels": {}, "masters": {}, "uuids": { "xvda1": [ "fe591198-9082-4b15-9b62-e83518524cd2" ] } }, "devices": { "xvda": { "holders": [], "host": "", "links": { "ids": [], "labels": [], "masters": [], "uuids": [] }, "model": null, "partitions": { "xvda1": { "holders": [], "links": { "ids": [], "labels": [], "masters": [], "uuids": [ "fe591198-9082-4b15-9b62-e83518524cd2" ] }, "sectors": "524285919", "sectorsize": 512, "size": "250.00 GB", "start": "2048", "uuid": "fe591198-9082-4b15-9b62-e83518524cd2" } }, "removable": "0", "rotational": "0", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "524288000", "sectorsize": "512", "size": "250.00 GB", "support_discard": "0", "vendor": null, "virtual": 1 } }, "discovered_interpreter_python": "/usr/libexec/platform-python", "distribution": "CentOS", "distribution_file_parsed": true, "distribution_file_path": "/etc/redhat-release", "distribution_file_variety": "RedHat", "distribution_major_version": "8", "distribution_release": "NA", "distribution_version": "8", "dns": { "nameservers": [ "10.29.169.13", "10.29.170.12", "10.2.32.1" ], "search": [ "testing-farm.us-east-1.aws.redhat.com" ] }, "domain": "", "effective_group_id": 0, "effective_user_id": 0, "env": { "BASH_FUNC_which%%": "() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "HOME": "/root", "LANG": "en_US.UTF-8", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "LOGNAME": "root", "LS_COLORS": "", "PATH": "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "PWD": "/root", "SELINUX_LEVEL_REQUESTED": "", "SELINUX_ROLE_REQUESTED": "", "SELINUX_USE_CURRENT_RANGE": "", "SHELL": "/bin/bash", "SHLVL": "2", "SSH_CLIENT": "10.31.47.219 41724 22", "SSH_CONNECTION": "10.31.47.219 41724 10.31.47.254 22", "SSH_TTY": "/dev/pts/0", "USER": "root", "XDG_RUNTIME_DIR": "/run/user/0", "XDG_SESSION_ID": "7", "_": "/usr/libexec/platform-python", "which_declare": "declare -f" }, "eth0": { "active": true, "device": "eth0", "features": { "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "on", "highdma": "off [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "off [fixed]", "netns_local": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "on [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_gro_list": "off", "rx_udp_gro_forwarding": "off", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "off [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "on", "tcp_segmentation_offload": "on", "tls_hw_record": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv4": "on [fixed]", "tx_checksum_ipv6": "on", "tx_checksum_sctp": "off [fixed]", "tx_checksumming": "on", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_list": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "on [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "off [fixed]", "tx_nocache_copy": "off", "tx_scatter_gather": "on", "tx_scatter_gather_fraglist": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_tcp6_segmentation": "on", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off", "tx_tcp_segmentation": "on", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "vlan_challenged": "off [fixed]" }, "hw_timestamp_filters": [], "ipv4": { "address": "10.31.47.254", "broadcast": "10.31.47.255", "netmask": "255.255.252.0", "network": "10.31.44.0" }, "ipv6": [ { "address": "fe80::6:8cff:fe5b:9bed", "prefix": "64", "scope": "link" } ], "macaddress": "02:06:8c:5b:9b:ed", "module": "xen_netfront", "mtu": 9001, "pciid": "vif-0", "promisc": false, "timestamping": [], "type": "ether" }, "fibre_channel_wwn": [], "fips": false, "form_factor": "Other", "fqdn": "managed-node3", "gather_subset": [ "all" ], "hostname": "managed-node3", "hostnqn": "", "interfaces": [ "lo", "eth0" ], "is_chroot": false, "iscsi_iqn": "", "kernel": "4.18.0-553.5.1.el8.x86_64", "kernel_version": "#1 SMP Tue May 21 05:46:01 UTC 2024", "lo": { "active": true, "device": "lo", "features": { "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "on", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "on [fixed]", "netns_local": "on [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "on [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_gro_list": "off", "rx_udp_gro_forwarding": "off", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "off [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "on", "tcp_segmentation_offload": "on", "tls_hw_record": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "tx_checksumming": "on", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_list": "on", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "on [fixed]", "tx_nocache_copy": "off [fixed]", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tx_sctp_segmentation": "on", "tx_tcp6_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp_segmentation": "on", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_udp_segmentation": "on", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "vlan_challenged": "on [fixed]" }, "hw_timestamp_filters": [], "ipv4": { "address": "127.0.0.1", "broadcast": "", "netmask": "255.0.0.0", "network": "127.0.0.0" }, "ipv6": [ { "address": "::1", "prefix": "128", "scope": "host" } ], "mtu": 65536, "promisc": false, "timestamping": [], "type": "loopback" }, "lsb": {}, "machine": "x86_64", "machine_id": "d3e3c495d99a450fafbc2a1a61acc098", "memfree_mb": 3004, "memory_mb": { "nocache": { "free": 3338, "used": 201 }, "real": { "free": 3004, "total": 3539, "used": 535 }, "swap": { "cached": 0, "free": 0, "total": 0, "used": 0 } }, "memtotal_mb": 3539, "module_setup": true, "mounts": [ { "block_available": 64386141, "block_size": 4096, "block_total": 65533179, "block_used": 1147038, "device": "/dev/xvda1", "fstype": "xfs", "inode_available": 130996213, "inode_total": 131071472, "inode_used": 75259, "mount": "/", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 263725633536, "size_total": 268423901184, "uuid": "fe591198-9082-4b15-9b62-e83518524cd2" } ], "nodename": "managed-node3", "os_family": "RedHat", "pkg_mgr": "dnf", "proc_cmdline": { "BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-553.5.1.el8.x86_64", "crashkernel": "auto", "net.ifnames": "0", "quiet": true, "rhgb": true, "ro": true, "root": "UUID=fe591198-9082-4b15-9b62-e83518524cd2" }, "processor": [ "0", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz", "1", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz" ], "processor_cores": 1, "processor_count": 1, "processor_threads_per_core": 2, "processor_vcpus": 2, "product_name": "HVM domU", "product_serial": "ec22a626-8bc9-9335-9697-4d236edaee8e", "product_uuid": "ec22a626-8bc9-9335-9697-4d236edaee8e", "product_version": "4.11.amazon", "python": { "executable": "/usr/libexec/platform-python", "has_sslcontext": true, "type": "cpython", "version": { "major": 3, "micro": 8, "minor": 6, "releaselevel": "final", "serial": 0 }, "version_info": [ 3, 6, 8, "final", 0 ] }, "python_version": "3.6.8", "real_group_id": 0, "real_user_id": 0, "selinux": { "config_mode": "enforcing", "mode": "enforcing", "policyvers": 33, "status": "enabled", "type": "targeted" }, "selinux_python_present": true, "service_mgr": "systemd", "ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAJmoPF1VcGnVGJiZReQ3FMq7KDjoo91kjJ1+jlyD04Fup6ocKX8L4c1Mk8hRRfooXMYts+WCWeb9QrgW1D+P3zqCSnTKryq+41bgnfu8BCASj1otowdm8KTAbxWBFppYwG5hyLJH6gDn8zNjf5H8y/OH8JoXMoU9C6YmtiiE48m/AAAAFQCmRu8T7V2Y0XDhYIHFoctcZNQ8rwAAAIBrWYsgmOxAH11coLFBmDkzULpDaQdDJbpp8NYJmJDG51NesGhXkaZq3uplOgkfRUeY6Hm9BzGkRwWbvoqmLk+tJ5h5uMsMGR5PZiAGAGOOV+TzaF1pRO+dU8altt2eRl4H7FA43zxNKu1wlJ+nL9w66F7KVP6XjwVSEmVjG53JrQAAAIACqGROkP0RAXQHVGazydVLFFNVkcQIX7n56kVSeg28IUrMPzL9tq+lNRMJe5zXo7hKLVae7szT18aZAWEXbZyUVAsm4ndrBhX/9WPemQ9cO9LLiYHYYmPqyaXMd/cispWFKly6z5hMG4Bd03Gs4L5GY6yr29EwYANdZbVX/sAJig==", "ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOD5g4FPBObQLfhVZ+V9NI7tUIi32VmIMkKEjl/3EyMK+/r0KjzytUNyuOImf/7b4MSg8EbzPKLDinK/TtfhZdA=", "ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIKjQjbgQH4GZBxVPx0C2mbj+mAXi433blujJ/r8gfwz/", "ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQDMOGTN+rw4KPPoxUFhimn4wmGIU+lt4RZG2REi4hZpvivrE7u580QWYeKknbTKawxW3E76x59zA/tfpxg3QG3GYakgE3DqYEMsy5xXPZ1t943T/XxThq8zTwpnirAxlIQw1GRjBYUnWwdgv15McSZRhZ6ij3IKyDspqR1LIoHZZPX/H74JG+TejvgwzyixEQxSOtwgHrhKcnrLDwfXeMMCNRdPU9sQkGaCSquVul81pTmox4sZ9EqdtOSydF+8ZxaBd1SexISeCHGR+tG8a7xLhNCvywIQ6N0MxVR+S1tX//ioluQzPX81wIwd1qIyxgKF7FMZ53bvH8UIl8mP0wb5/cWP9JGR7rMsT16kP8RL9+7qoWp3c4q5mrv+VU2TxkYm5BSwoex1sxzZcS7XXb5SFx0yGJWs4YmgR/HO47PObAYBs76Gr6mewYCAfsGX7pK6sN/GjN2MtZrWr5MsNde9SF9Aq4U4f+YVB+ipPGrWw8Rkydx6Cybdt0CpMzPSO4k=", "swapfree_mb": 0, "swaptotal_mb": 0, "system": "Linux", "system_capabilities": [], "system_capabilities_enforced": "False", "system_vendor": "Xen", "uptime_seconds": 193, "user_dir": "/root", "user_gecos": "root", "user_gid": 0, "user_id": "root", "user_shell": "/bin/bash", "user_uid": 0, "userspace_architecture": "x86_64", "userspace_bits": "64", "virtualization_role": "guest", "virtualization_type": "xen" } TASK [Remove NTP providers] **************************************************** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tests_ntp_provider1.yml:20 Monday 02 February 2026 18:19:20 -0500 (0:00:00.018) 0:00:16.833 ******* changed: [managed-node3] => (item=chrony) => { "ansible_loop_var": "item", "changed": true, "item": "chrony", "rc": 0, "results": [ "Removed: chrony-4.5-1.el8.x86_64" ] } ok: [managed-node3] => (item=ntp) => { "ansible_loop_var": "item", "changed": false, "item": "ntp", "rc": 0, "results": [] } MSG: Nothing to do META: ran handlers TASK [fedora.linux_system_roles.timesync : Set version specific variables] ***** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:2 Monday 02 February 2026 18:19:26 -0500 (0:00:05.792) 0:00:22.625 ******* included: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.timesync : Ensure ansible_facts used by role] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:2 Monday 02 February 2026 18:19:26 -0500 (0:00:00.027) 0:00:22.653 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Check if system is ostree] ********** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:10 Monday 02 February 2026 18:19:26 -0500 (0:00:00.022) 0:00:22.675 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Set flag to indicate system is ostree] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:15 Monday 02 February 2026 18:19:26 -0500 (0:00:00.014) 0:00:22.690 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:22 Monday 02 February 2026 18:19:26 -0500 (0:00:00.018) 0:00:22.708 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Set flag if transactional-update exists] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:27 Monday 02 February 2026 18:19:26 -0500 (0:00:00.014) 0:00:22.722 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Set platform/version specific variables] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:31 Monday 02 February 2026 18:19:26 -0500 (0:00:00.014) 0:00:22.737 ******* ok: [managed-node3] => { "ansible_facts": { "timesync_chrony_conf_path": "/etc/chrony.conf", "timesync_chrony_dhcp_sourcedir": "", "timesync_chrony_sysconfig_options": "", "timesync_chrony_sysconfig_path": "/etc/sysconfig/chronyd", "timesync_ntp_provider_os_default": "chrony", "timesync_ntp_sysconfig_path": "/etc/sysconfig/ntpd", "timesync_phc2sys_sysconfig_path": "/etc/sysconfig/phc2sys", "timesync_ptp4l_sysconfig_path": "/etc/sysconfig/ptp4l", "timesync_timemaster_config_path": "/etc/timemaster.conf" }, "ansible_included_var_files": [ "/tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/vars/CentOS_8.yml" ], "changed": false } TASK [fedora.linux_system_roles.timesync : Run systemctl] ********************** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:47 Monday 02 February 2026 18:19:26 -0500 (0:00:00.026) 0:00:22.763 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Require installed systemd] ********** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:55 Monday 02 February 2026 18:19:26 -0500 (0:00:00.014) 0:00:22.778 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Set flag to indicate that systemd runtime operations are available] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:60 Monday 02 February 2026 18:19:26 -0500 (0:00:00.014) 0:00:22.792 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Populate service facts] ************* task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:5 Monday 02 February 2026 18:19:26 -0500 (0:00:00.016) 0:00:22.809 ******* ok: [managed-node3] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "active", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "masked" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "tcsd.service": { "name": "tcsd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "timedatex.service": { "name": "timedatex.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [fedora.linux_system_roles.timesync : Set variable `timesync_services` with filtered uniq service names] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:9 Monday 02 February 2026 18:19:27 -0500 (0:00:01.532) 0:00:24.341 ******* ok: [managed-node3] => { "ansible_facts": { "timesync_services": [ "dracut-shutdown-onfailure", "initrd-udevadm-cleanup-db", "nfs-idmapd", "systemd-logind", "rescue", "user@0", "initrd-cleanup", "dracut-pre-mount", "sssd-autofs", "sssd-kcm", "man-db-restart-cache-update", "sshd-keygen@ecdsa", "systemd-tmpfiles-setup", "sssd-sudo", "polkit", "dracut-mount", "selinux-autorelabel-mark", "auditd", "kvm_stat", "sshd-keygen@rsa", "systemd-hibernate-resume", "user-runtime-dir@0", "systemd-firstboot", "systemd-quotacheck", "systemd-ask-password-console", "getty@tty1", "kmod-static-nodes", "plymouth-reboot", "systemd-coredump", "user-runtime-dir", "rpc-statd-notify", "nfs-mountd", "systemd-journal-flush", "systemd-udev-trigger", "plymouth-kexec", "sssd", "systemd-tmpfiles-setup-dev", "rpc-statd", "fstrim", "systemd-modules-load", "rdisc", "sshd-keygen", "systemd-hostnamed", "initrd-switch-root", "kdump", "systemd-hybrid-sleep", "dracut-pre-trigger", "iprinit", "nfsdcld", "dbus-org.freedesktop.portable1", "systemd-sysusers", "systemd-udevd", "dnf-system-upgrade-cleanup", "getty", "nfs-convert", "systemd-poweroff", "systemd-backlight", "cloud-init-local", "systemd-portabled", "systemd-update-done", "systemd-suspend", "systemd-random-seed", "serial-getty", "microcode", "systemd-fsck-root", "plymouth-read-write", "messagebus", "systemd-volatile-root", "systemd-ask-password-wall", "systemd-journal-catalog-update", "systemd-resolved", "cpupower", "cloud-init-hotplugd", "nfs-utils", "systemd-udev-settle", "dracut-pre-udev", "systemd-vconsole-setup", "dbus-org.freedesktop.nm-dispatcher", "rsyslog", "man-db-cache-update", "dbus-org.freedesktop.hostname1", "systemd-fsck", "system-update-cleanup", "systemd-binfmt", "quotaon", "NetworkManager-wait-online", "NetworkManager", "dbus", "systemd-hwdb-update", "dnf-system-upgrade", "sssd-ssh", "cloud-config", "rpc-gssd", "sssd-nss", "nftables", "import-state", "systemd-journald", "sssd-pac", "ldconfig", "plymouth-switch-root", "systemd-update-utmp-runlevel", "systemd-user-sessions", "sssd-pam", "nis-domainname", "restraintd", "user", "oddjobd", "auth-rpcgss-module", "container-getty", "dracut-pre-pivot", "ebtables", "plymouth-halt", "plymouth-poweroff", "dbus-org.freedesktop.locale1", "iprdump", "dracut-initqueue", "systemd-update-utmp", "rpcbind", "unbound-anchor", "systemd-initctl", "systemd-hibernate", "nfs-server", "systemd-reboot", "systemd-kexec", "systemd-sysctl", "halt-local", "systemd-suspend-then-hibernate", "systemd-ask-password-plymouth", "qemu-guest-agent", "sshd", "gssproxy", "grub-boot-indeterminate", "cloud-final", "selinux-autorelabel", "systemd-remount-fs", "iprupdate", "emergency", "dnf-makecache", "irqbalance", "loadmodules", "plymouth-switch-root-initramfs", "tcsd", "rngd", "plymouth-quit-wait", "rc-local", "timedatex", "debug-shell", "plymouth-start", "systemd-exit", "systemd-machine-id-commit", "plymouth-quit", "firewalld", "tuned", "dracut-shutdown", "autovt", "systemd-pstore", "systemd-halt", "teamd", "systemd-rfkill", "cloud-init", "nfs-blkmap", "systemd-tmpfiles-clean", "syslog", "dracut-cmdline", "systemd-localed", "crond", "dbus-org.freedesktop.timedate1", "initrd-parse-etc", "console-getty", "NetworkManager-dispatcher", "sshd-keygen@ed25519", "dbus-org.freedesktop.login1" ] }, "changed": false } TASK [fedora.linux_system_roles.timesync : Populate service facts when in bootc mode] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:14 Monday 02 February 2026 18:19:27 -0500 (0:00:00.033) 0:00:24.375 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Set variable `timesync_services` with filtered uniq service names when in bootc mode] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:20 Monday 02 February 2026 18:19:27 -0500 (0:00:00.015) 0:00:24.390 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Check that variable 'timesync_services' is defined] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:25 Monday 02 February 2026 18:19:27 -0500 (0:00:00.015) 0:00:24.405 ******* ok: [managed-node3] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.timesync : Check if only NTP is needed] ******** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:30 Monday 02 February 2026 18:19:28 -0500 (0:00:00.019) 0:00:24.425 ******* ok: [managed-node3] => { "ansible_facts": { "timesync_mode": 1 }, "changed": false } TASK [fedora.linux_system_roles.timesync : Check if single PTP is needed] ****** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:35 Monday 02 February 2026 18:19:28 -0500 (0:00:00.017) 0:00:24.442 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Check if both NTP and PTP are needed] *** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:44 Monday 02 February 2026 18:19:28 -0500 (0:00:00.015) 0:00:24.458 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.timesync : Determine current NTP provider] ***** task path: /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:49 Monday 02 February 2026 18:19:28 -0500 (0:00:00.014) 0:00:24.472 ******* fatal: [managed-node3]: FAILED! => {} MSG: module (fedora.linux_system_roles.timesync_provider) is missing interpreter line PLAY RECAP ********************************************************************* managed-node3 : ok=18 changed=1 unreachable=0 failed=1 skipped=15 rescued=0 ignored=0 SYSTEM ROLES ERRORS BEGIN v1 [ { "ansible_version": "2.9.27", "end_time": "2026-02-02T23:19:28.157823+00:00Z", "host": "managed-node3", "message": "module (fedora.linux_system_roles.timesync_provider) is missing interpreter line", "start_time": "2026-02-02T23:19:28.053171+00:00Z", "task_name": "Determine current NTP provider", "task_path": "/tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:49" } ] SYSTEM ROLES ERRORS END v1 TASKS RECAP ******************************************************************** Monday 02 February 2026 18:19:28 -0500 (0:00:00.107) 0:00:24.580 ******* =============================================================================== Ensure iproute for gathering default_ipv4 fact ------------------------- 14.13s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tasks/setup.yml:15 Remove NTP providers ---------------------------------------------------- 5.79s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tests_ntp_provider1.yml:20 fedora.linux_system_roles.timesync : Populate service facts ------------- 1.53s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:5 Gathering Facts --------------------------------------------------------- 0.87s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tests_ntp_provider1.yml:2 fedora.linux_system_roles.timesync : Check if system is ostree ---------- 0.55s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:10 fedora.linux_system_roles.timesync : Run systemctl ---------------------- 0.44s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:47 fedora.linux_system_roles.timesync : Check if transactional-update exists in /sbin --- 0.37s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:22 fedora.linux_system_roles.timesync : Determine current NTP provider ----- 0.11s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:49 fedora.linux_system_roles.timesync : Set flag to indicate that systemd runtime operations are available --- 0.09s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:60 fedora.linux_system_roles.timesync : Require installed systemd ---------- 0.06s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:55 fedora.linux_system_roles.timesync : Ensure ansible_facts used by role --- 0.05s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:2 fedora.linux_system_roles.timesync : Set platform/version specific variables --- 0.04s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:31 fedora.linux_system_roles.timesync : Set flag to indicate system is ostree --- 0.04s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:15 fedora.linux_system_roles.timesync : Set flag if transactional-update exists --- 0.04s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:27 Run the role only to get vars needed for validation --------------------- 0.04s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tasks/setup.yml:3 fedora.linux_system_roles.timesync : Set variable `timesync_services` with filtered uniq service names --- 0.03s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:9 Ensure ansible_facts used by test --------------------------------------- 0.03s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/tests/timesync/tasks/setup.yml:21 fedora.linux_system_roles.timesync : Set version specific variables ----- 0.03s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/main.yml:2 fedora.linux_system_roles.timesync : Set platform/version specific variables --- 0.03s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:31 fedora.linux_system_roles.timesync : Ensure ansible_facts used by role --- 0.02s /tmp/collections-Mpd/ansible_collections/fedora/linux_system_roles/roles/timesync/tasks/set_vars.yml:2 -- Logs begin at Mon 2026-02-02 18:15:51 EST, end at Mon 2026-02-02 18:19:28 EST. -- Feb 02 18:19:02 managed-node3 sshd[7991]: Accepted publickey for root from 10.31.47.219 port 41732 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Feb 02 18:19:02 managed-node3 systemd[1]: Started Session 8 of user root. -- Subject: Unit session-8.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-8.scope has finished starting up. -- -- The start-up result is done. Feb 02 18:19:02 managed-node3 systemd-logind[620]: New session 8 of user root. -- Subject: A new session 8 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 8 has been created for the user root. -- -- The leading process of the session is 7991. Feb 02 18:19:02 managed-node3 sshd[7991]: pam_unix(sshd:session): session opened for user root by (uid=0) Feb 02 18:19:02 managed-node3 sshd[7994]: Received disconnect from 10.31.47.219 port 41732:11: disconnected by user Feb 02 18:19:02 managed-node3 sshd[7994]: Disconnected from user root 10.31.47.219 port 41732 Feb 02 18:19:02 managed-node3 sshd[7991]: pam_unix(sshd:session): session closed for user root Feb 02 18:19:02 managed-node3 systemd[1]: session-8.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit session-8.scope has successfully entered the 'dead' state. Feb 02 18:19:02 managed-node3 systemd-logind[620]: Session 8 logged out. Waiting for processes to exit. Feb 02 18:19:02 managed-node3 systemd-logind[620]: Removed session 8. -- Subject: Session 8 has been terminated -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 8 has been terminated. Feb 02 18:19:02 managed-node3 sshd[8015]: Accepted publickey for root from 10.31.47.219 port 41748 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Feb 02 18:19:02 managed-node3 systemd[1]: Started Session 9 of user root. -- Subject: Unit session-9.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-9.scope has finished starting up. -- -- The start-up result is done. Feb 02 18:19:02 managed-node3 systemd-logind[620]: New session 9 of user root. -- Subject: A new session 9 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 9 has been created for the user root. -- -- The leading process of the session is 8015. Feb 02 18:19:02 managed-node3 sshd[8015]: pam_unix(sshd:session): session opened for user root by (uid=0) Feb 02 18:19:02 managed-node3 sshd[8018]: Received disconnect from 10.31.47.219 port 41748:11: disconnected by user Feb 02 18:19:02 managed-node3 sshd[8018]: Disconnected from user root 10.31.47.219 port 41748 Feb 02 18:19:02 managed-node3 sshd[8015]: pam_unix(sshd:session): session closed for user root Feb 02 18:19:02 managed-node3 systemd[1]: session-9.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit session-9.scope has successfully entered the 'dead' state. Feb 02 18:19:02 managed-node3 systemd-logind[620]: Session 9 logged out. Waiting for processes to exit. Feb 02 18:19:02 managed-node3 systemd-logind[620]: Removed session 9. -- Subject: Session 9 has been terminated -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 9 has been terminated. Feb 02 18:19:04 managed-node3 platform-python[8180]: ansible-setup Invoked with gather_subset=['all'] gather_timeout=10 filter=* fact_path=/etc/ansible/facts.d Feb 02 18:19:05 managed-node3 platform-python[8328]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 02 18:19:05 managed-node3 platform-python[8451]: ansible-stat Invoked with path=/sbin/transactional-update follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 02 18:19:06 managed-node3 platform-python[8574]: ansible-command Invoked with _raw_params=systemctl is-system-running warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 02 18:19:06 managed-node3 platform-python[8698]: ansible-dnf Invoked with name=['iproute'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 02 18:19:20 managed-node3 platform-python[8869]: ansible-dnf Invoked with name=['chrony'] state=absent allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 02 18:19:23 managed-node3 chronyd[629]: chronyd exiting Feb 02 18:19:23 managed-node3 systemd[1]: Stopping NTP client/server... -- Subject: Unit chronyd.service has begun shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit chronyd.service has begun shutting down. Feb 02 18:19:23 managed-node3 systemd[1]: chronyd.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit chronyd.service has successfully entered the 'dead' state. Feb 02 18:19:23 managed-node3 systemd[1]: Stopped NTP client/server. -- Subject: Unit chronyd.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit chronyd.service has finished shutting down. Feb 02 18:19:23 managed-node3 systemd[1]: Reloading. Feb 02 18:19:23 managed-node3 systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. -- Subject: Unit run-r1ac3f2caa0df4baab848365938a2ca3e.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit run-r1ac3f2caa0df4baab848365938a2ca3e.service has finished starting up. -- -- The start-up result is done. Feb 02 18:19:23 managed-node3 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 02 18:19:23 managed-node3 systemd[1]: Starting man-db-cache-update.service... -- Subject: Unit man-db-cache-update.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has begun starting up. Feb 02 18:19:24 managed-node3 systemd[1]: man-db-cache-update.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 02 18:19:24 managed-node3 systemd[1]: Started man-db-cache-update.service. -- Subject: Unit man-db-cache-update.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has finished starting up. -- -- The start-up result is done. Feb 02 18:19:24 managed-node3 systemd[1]: run-r1ac3f2caa0df4baab848365938a2ca3e.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-r1ac3f2caa0df4baab848365938a2ca3e.service has successfully entered the 'dead' state. Feb 02 18:19:24 managed-node3 platform-python[9354]: ansible-dnf Invoked with name=['ntp'] state=absent allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 02 18:19:26 managed-node3 platform-python[9481]: ansible-service_facts Invoked Feb 02 18:19:28 managed-node3 sshd[9644]: Accepted publickey for root from 10.31.47.219 port 41084 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Feb 02 18:19:28 managed-node3 systemd[1]: Started Session 10 of user root. -- Subject: Unit session-10.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-10.scope has finished starting up. -- -- The start-up result is done. Feb 02 18:19:28 managed-node3 systemd-logind[620]: New session 10 of user root. -- Subject: A new session 10 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 10 has been created for the user root. -- -- The leading process of the session is 9644. Feb 02 18:19:28 managed-node3 sshd[9644]: pam_unix(sshd:session): session opened for user root by (uid=0) Feb 02 18:19:28 managed-node3 sshd[9647]: Received disconnect from 10.31.47.219 port 41084:11: disconnected by user Feb 02 18:19:28 managed-node3 sshd[9647]: Disconnected from user root 10.31.47.219 port 41084 Feb 02 18:19:28 managed-node3 sshd[9644]: pam_unix(sshd:session): session closed for user root Feb 02 18:19:28 managed-node3 systemd[1]: session-10.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit session-10.scope has successfully entered the 'dead' state. Feb 02 18:19:28 managed-node3 systemd-logind[620]: Session 10 logged out. Waiting for processes to exit. Feb 02 18:19:28 managed-node3 systemd-logind[620]: Removed session 10. -- Subject: Session 10 has been terminated -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 10 has been terminated. Feb 02 18:19:28 managed-node3 sshd[9668]: Accepted publickey for root from 10.31.47.219 port 41086 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Feb 02 18:19:28 managed-node3 systemd[1]: Started Session 11 of user root. -- Subject: Unit session-11.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-11.scope has finished starting up. -- -- The start-up result is done. Feb 02 18:19:28 managed-node3 systemd-logind[620]: New session 11 of user root. -- Subject: A new session 11 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 11 has been created for the user root. -- -- The leading process of the session is 9668. Feb 02 18:19:28 managed-node3 sshd[9668]: pam_unix(sshd:session): session opened for user root by (uid=0)