ansible-playbook [core 2.16.18] config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.12/site-packages/ansible ansible collection location = /tmp/collections-LJi executable location = /usr/local/bin/ansible-playbook python version = 3.12.1 (main, Feb 21 2024, 14:18:26) [GCC 8.5.0 20210514 (Red Hat 8.5.0-21)] (/usr/bin/python3.12) jinja version = 3.1.6 libyaml = True No config file found; using defaults running playbook inside collection fedora.linux_system_roles Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'jsonl', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_purge_rules.yml ************************************************ 1 plays in /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml PLAY [Test auditd_purge_rules purge and idempotence] *************************** TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:5 Tuesday 21 April 2026 21:26:18 -0400 (0:00:00.019) 0:00:00.019 ********* ok: [managed-node1] TASK [Back up existing auditd configuration and rules metadata] **************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:13 Tuesday 21 April 2026 21:26:19 -0400 (0:00:01.091) 0:00:01.111 ********* included: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/setup.yml for managed-node1 TASK [Check for existing auditd.conf] ****************************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/setup.yml:7 Tuesday 21 April 2026 21:26:19 -0400 (0:00:00.067) 0:00:01.179 ********* ok: [managed-node1] => { "changed": false, "stat": { "atime": 1776821144.675373, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "33eefb6af1c751110a125b8b468118e6f0edb244", "ctime": 1776821140.7013352, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 119537858, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0640", "mtime": 1776821140.3443317, "nlink": 1, "path": "/etc/audit/auditd.conf", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 852, "uid": 0, "version": "3051320685", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Check for existing custom.rules] ***************************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/setup.yml:12 Tuesday 21 April 2026 21:26:19 -0400 (0:00:00.437) 0:00:01.616 ********* ok: [managed-node1] => { "changed": false, "stat": { "atime": 1776821143.3643606, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "beaedc34af6f74d852b504550a00030393947a4b", "ctime": 1776821141.8453462, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 130023618, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0664", "mtime": 1776821141.5933437, "nlink": 1, "path": "/etc/audit/rules.d/custom.rules", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 293, "uid": 0, "version": "2419061716", "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Remember whether custom.rules existed] *********************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/setup.yml:17 Tuesday 21 April 2026 21:26:19 -0400 (0:00:00.332) 0:00:01.948 ********* ok: [managed-node1] => { "ansible_facts": { "__auditd_integration_had_custom_rules": true }, "changed": false } TASK [Back up auditd.conf before test] ***************************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/setup.yml:21 Tuesday 21 April 2026 21:26:19 -0400 (0:00:00.015) 0:00:01.964 ********* changed: [managed-node1] => { "changed": true, "checksum": "33eefb6af1c751110a125b8b468118e6f0edb244", "dest": "/root/.lsr_auditd_integration_auditd.conf.bak", "gid": 0, "group": "root", "md5sum": "eda8c45ce09c30766469af2714e2eff9", "mode": "0640", "owner": "root", "secontext": "system_u:object_r:admin_home_t:s0", "size": 852, "src": "/etc/audit/auditd.conf", "state": "file", "uid": 0 } TASK [Back up custom.rules before test] **************************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/setup.yml:29 Tuesday 21 April 2026 21:26:20 -0400 (0:00:00.447) 0:00:02.411 ********* changed: [managed-node1] => { "changed": true, "checksum": "beaedc34af6f74d852b504550a00030393947a4b", "dest": "/root/.lsr_auditd_integration_custom.rules.bak", "gid": 0, "group": "root", "md5sum": "02a93488fb84ceabb705a049f70ced6b", "mode": "0664", "owner": "root", "secontext": "system_u:object_r:admin_home_t:s0", "size": 293, "src": "/etc/audit/rules.d/custom.rules", "state": "file", "uid": 0 } TASK [Record auditd.conf backup path for cleanup] ****************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/setup.yml:37 Tuesday 21 April 2026 21:26:20 -0400 (0:00:00.340) 0:00:02.752 ********* ok: [managed-node1] => { "ansible_facts": { "__auditd_integration_backup_auditd_conf": "/root/.lsr_auditd_integration_auditd.conf.bak" }, "changed": false } TASK [Record custom.rules backup path for cleanup] ***************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/setup.yml:42 Tuesday 21 April 2026 21:26:20 -0400 (0:00:00.016) 0:00:02.768 ********* ok: [managed-node1] => { "ansible_facts": { "__auditd_integration_backup_custom_rules": "/root/.lsr_auditd_integration_custom.rules.bak" }, "changed": false } TASK [Ensure rules.d backup parent directory is absent] ************************ task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:16 Tuesday 21 April 2026 21:26:20 -0400 (0:00:00.017) 0:00:02.785 ********* ok: [managed-node1] => { "changed": false, "path": "/root/.lsr_auditd_purge_rulesd.backup", "state": "absent" } TASK [Create parent directory for rules.d snapshot] **************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:21 Tuesday 21 April 2026 21:26:21 -0400 (0:00:00.441) 0:00:03.226 ********* changed: [managed-node1] => { "changed": true, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/root/.lsr_auditd_purge_rulesd.backup", "secontext": "unconfined_u:object_r:admin_home_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [Snapshot rules.d tree for full restore after test] *********************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:27 Tuesday 21 April 2026 21:26:21 -0400 (0:00:00.336) 0:00:03.563 ********* changed: [managed-node1] => { "changed": true, "checksum": null, "dest": "/root/.lsr_auditd_purge_rulesd.backup/", "gid": 0, "group": "root", "md5sum": null, "mode": "0750", "owner": "root", "secontext": "unconfined_u:object_r:admin_home_t:s0", "size": 45, "src": "/etc/audit/rules.d/", "state": "directory", "uid": 0 } TASK [Install decoy rules fragment to be removed by purge] ********************* task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:34 Tuesday 21 April 2026 21:26:21 -0400 (0:00:00.327) 0:00:03.891 ********* changed: [managed-node1] => { "changed": true, "checksum": "35324fa08aa315fc1ab79570382f077fec44dce9", "dest": "/etc/audit/rules.d/99-lsr-purge-decoy.rules", "gid": 0, "group": "root", "md5sum": "f628c72c0861f587eff2d10f92b8e21b", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:auditd_etc_t:s0", "size": 89, "src": "/root/.ansible/tmp/ansible-tmp-1776821181.9469764-9982-220291465627703/source", "state": "file", "uid": 0 } TASK [Run auditd role with purge enabled] ************************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:42 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.620) 0:00:04.511 ********* included: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/run_role_with_clear_facts.yml for managed-node1 TASK [Clear facts] ************************************************************* task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/run_role_with_clear_facts.yml:9 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.021) 0:00:04.533 ********* META: facts cleared TASK [Run the role] ************************************************************ task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/run_role_with_clear_facts.yml:23 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.000) 0:00:04.534 ********* skipping: [managed-node1] => { "changed": false, "false_condition": "__sr_failed_when is defined", "skip_reason": "Conditional result was False" } TASK [Run the role normally] *************************************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/run_role_with_clear_facts.yml:33 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.018) 0:00:04.552 ********* TASK [fedora.linux_system_roles.auditd : Validating arguments against arg spec 'main' - Manage auditd and audit rules] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:5 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.079) 0:00:04.631 ********* ok: [managed-node1] => { "changed": false, "validate_args_context": { "argument_spec_name": "main", "name": "auditd", "path": "/tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd", "type": "role" } } MSG: The arg spec validation passed TASK [fedora.linux_system_roles.auditd : Validate role parameters] ************* task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/main.yml:3 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.016) 0:00:04.648 ********* included: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml for managed-node1 TASK [fedora.linux_system_roles.auditd : Assert num_logs range (num_logs_parser)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:4 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.031) 0:00:04.679 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert freq range (freq_parser)] ****** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:11 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.018) 0:00:04.698 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert incremental flush requires non-zero freq (sanity_check)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:18 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.018) 0:00:04.716 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert priority_boost range (priority_boost_parser)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:25 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.018) 0:00:04.735 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert q_depth range (q_depth_parser)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:31 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.018) 0:00:04.754 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert auditd_maximum_rate is null or a non-negative integer] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:38 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.019) 0:00:04.774 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert auditd_backlog_wait_time is null or a non-negative integer] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:48 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.024) 0:00:04.798 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert max_restarts range (max_restarts_parser)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:58 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.024) 0:00:04.822 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert tcp_listen_port range when listener enabled in build (tcp_listen_port_parser)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:64 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.018) 0:00:04.841 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert tcp_listen_queue range (tcp_listen_queue_parser)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:70 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.018) 0:00:04.860 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert tcp_max_per_addr range (tcp_max_per_addr_parser)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:76 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.019) 0:00:04.879 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert tcp_client_max_idle range (tcp_client_max_idle_parser)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:82 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.020) 0:00:04.899 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert tcp_client_ports format (tcp_client_ports_parser)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:88 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.018) 0:00:04.918 ********* skipping: [managed-node1] => { "changed": false, "false_condition": "auditd_tcp_client_ports is string", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.auditd : Assert tcp_client_ports range order] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:98 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.016) 0:00:04.935 ********* skipping: [managed-node1] => { "changed": false, "false_condition": "auditd_tcp_client_ports is string", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.auditd : Assert space_left_action rejects halt (space_action_parser)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:107 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.017) 0:00:04.952 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert space_left percentage is between 1 and 99 when given as N%] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:113 Tuesday 21 April 2026 21:26:22 -0400 (0:00:00.017) 0:00:04.970 ********* skipping: [managed-node1] => { "changed": false, "false_condition": "(auditd_space_left | string) is match('^[0-9]+%$')", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.auditd : Assert admin_space_left percentage is between 1 and 99 when given as N%] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:125 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.018) 0:00:04.989 ********* skipping: [managed-node1] => { "changed": false, "false_condition": "(auditd_admin_space_left | string) is match('^[0-9]+%$')", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.auditd : Assert space_left is greater than admin_space_left when both use same form] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:137 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.018) 0:00:05.007 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert disk_full_action rejects email (disk_full_action_parser)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:161 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.026) 0:00:05.034 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert disk_error_action rejects email and rotate (disk_error_action_parser)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:167 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.016) 0:00:05.051 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert exec companion paths when action is exec] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:173 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.017) 0:00:05.069 ********* skipping: [managed-node1] => { "changed": false, "false_condition": "auditd_max_log_file_action | lower == 'exec'", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.auditd : Assert space_left_action exec path] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:180 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.014) 0:00:05.083 ********* skipping: [managed-node1] => { "changed": false, "false_condition": "auditd_space_left_action | lower == 'exec'", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.auditd : Assert admin_space_left_action exec path] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:187 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.014) 0:00:05.098 ********* skipping: [managed-node1] => { "changed": false, "false_condition": "auditd_admin_space_left_action | lower == 'exec'", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.auditd : Assert disk_full_action exec path] **** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:194 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.014) 0:00:05.112 ********* skipping: [managed-node1] => { "changed": false, "false_condition": "auditd_disk_full_action | lower == 'exec'", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.auditd : Assert disk_error_action exec path] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:201 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.013) 0:00:05.126 ********* skipping: [managed-node1] => { "changed": false, "false_condition": "auditd_disk_error_action | lower == 'exec'", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.auditd : Assert name when name_format is user (resolve_node)] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:208 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.014) 0:00:05.140 ********* skipping: [managed-node1] => { "changed": false, "false_condition": "auditd_name_format | lower == 'user'", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.auditd : Validate auditd_rules structure and values] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_role_vars.yml:215 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.014) 0:00:05.154 ********* included: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml for managed-node1 TASK [fedora.linux_system_roles.auditd : Assert auditd_rules is a list of dicts] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml:4 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.041) 0:00:05.196 ********* ok: [managed-node1] => { "changed": false } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert each auditd_rules entry has required action and filter] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml:13 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.027) 0:00:05.223 ********* ok: [managed-node1] => (item=always,exit) => { "ansible_loop_var": "item", "changed": false, "item": { "action": "always", "arch": "b64", "filter": "exit", "keyname": "lsr_auditd_purge_test", "path": "/etc/issue", "permission": [ "read" ], "syscall": "open" } } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert auditd_rules entries use only supported keys] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml:29 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.032) 0:00:05.256 ********* ok: [managed-node1] => (item={'action': 'always', 'filter': 'exit', 'arch': 'b64', 'syscall': 'open', 'path': '/etc/issue', 'permission': ['read'], 'keyname': 'lsr_auditd_purge_test'}) => { "ansible_loop_var": "item", "changed": false, "item": { "action": "always", "arch": "b64", "filter": "exit", "keyname": "lsr_auditd_purge_test", "path": "/etc/issue", "permission": [ "read" ], "syscall": "open" } } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert auditd_rules path and dir are mutually exclusive] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml:49 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.026) 0:00:05.283 ********* ok: [managed-node1] => (item={'action': 'always', 'filter': 'exit', 'arch': 'b64', 'syscall': 'open', 'path': '/etc/issue', 'permission': ['read'], 'keyname': 'lsr_auditd_purge_test'}) => { "ansible_loop_var": "item", "changed": false, "item": { "action": "always", "arch": "b64", "filter": "exit", "keyname": "lsr_auditd_purge_test", "path": "/etc/issue", "permission": [ "read" ], "syscall": "open" } } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert auditd_rules path or dir requires filter exit] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml:56 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.024) 0:00:05.307 ********* ok: [managed-node1] => (item={'action': 'always', 'filter': 'exit', 'arch': 'b64', 'syscall': 'open', 'path': '/etc/issue', 'permission': ['read'], 'keyname': 'lsr_auditd_purge_test'}) => { "ansible_loop_var": "item", "changed": false, "item": { "action": "always", "arch": "b64", "filter": "exit", "keyname": "lsr_auditd_purge_test", "path": "/etc/issue", "permission": [ "read" ], "syscall": "open" } } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert auditd_rules syscall requires filter exit or io_uring] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml:64 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.025) 0:00:05.333 ********* ok: [managed-node1] => (item={'action': 'always', 'filter': 'exit', 'arch': 'b64', 'syscall': 'open', 'path': '/etc/issue', 'permission': ['read'], 'keyname': 'lsr_auditd_purge_test'}) => { "ansible_loop_var": "item", "changed": false, "item": { "action": "always", "arch": "b64", "filter": "exit", "keyname": "lsr_auditd_purge_test", "path": "/etc/issue", "permission": [ "read" ], "syscall": "open" } } MSG: All assertions passed TASK [fedora.linux_system_roles.auditd : Assert auditd_rules field is a non-empty string or list of non-empty strings] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml:72 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.024) 0:00:05.357 ********* skipping: [managed-node1] => (item={'action': 'always', 'filter': 'exit', 'arch': 'b64', 'syscall': 'open', 'path': '/etc/issue', 'permission': ['read'], 'keyname': 'lsr_auditd_purge_test'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "item.field is defined", "item": { "action": "always", "arch": "b64", "filter": "exit", "keyname": "lsr_auditd_purge_test", "path": "/etc/issue", "permission": [ "read" ], "syscall": "open" }, "skip_reason": "Conditional result was False" } skipping: [managed-node1] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.auditd : Assert auditd_rules filesystem filter includes fstype=debugfs or fstype=tracefs] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml:82 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.021) 0:00:05.378 ********* skipping: [managed-node1] => (item={'action': 'always', 'filter': 'exit', 'arch': 'b64', 'syscall': 'open', 'path': '/etc/issue', 'permission': ['read'], 'keyname': 'lsr_auditd_purge_test'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "item.filter == 'filesystem'", "item": { "action": "always", "arch": "b64", "filter": "exit", "keyname": "lsr_auditd_purge_test", "path": "/etc/issue", "permission": [ "read" ], "syscall": "open" }, "skip_reason": "Conditional result was False" } skipping: [managed-node1] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.auditd : Assert auditd_rules field fragments use a supported operator] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml:104 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.016) 0:00:05.395 ********* skipping: [managed-node1] => (item={'action': 'always', 'filter': 'exit', 'arch': 'b64', 'syscall': 'open', 'path': '/etc/issue', 'permission': ['read'], 'keyname': 'lsr_auditd_purge_test'}) => { "ansible_loop_var": "item", "changed": false, "false_condition": "item.field is defined", "item": { "action": "always", "arch": "b64", "filter": "exit", "keyname": "lsr_auditd_purge_test", "path": "/etc/issue", "permission": [ "read" ], "syscall": "open" }, "skip_reason": "Conditional result was False" } skipping: [managed-node1] => { "changed": false } MSG: All items skipped TASK [fedora.linux_system_roles.auditd : Assert auditd_rules permission is a non-empty string or list of non-empty strings] *** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml:120 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.021) 0:00:05.416 ********* fatal: [managed-node1]: FAILED! => {} MSG: The conditional check 'item.permission is auditd_non_empty_str_or_list' failed. The error was: template error while templating string: Could not load "auditd_non_empty_str_or_list": 'auditd_non_empty_str_or_list'. String: {% set __cres = item.permission is auditd_non_empty_str_or_list %}{{ [true if __cres else false, __cres.__class__.__name__] }}. Could not load "auditd_non_empty_str_or_list": 'auditd_non_empty_str_or_list' The error appears to be in '/tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml': line 123, column 9, but may be elsewhere in the file depending on the exact syntax problem. The offending line appears to be: that: - item.permission is auditd_non_empty_str_or_list ^ here TASK [Check whether purge test rules.d snapshot exists] ************************ task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:135 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.025) 0:00:05.441 ********* ok: [managed-node1] => { "ansible_facts": { "discovered_interpreter_python": "/usr/libexec/platform-python" }, "changed": false, "stat": { "atime": 1776821181.864703, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1776821181.867703, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 226492608, "isblk": false, "ischr": false, "isdir": true, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/directory", "mode": "0750", "mtime": 1776821181.864703, "nlink": 2, "path": "/root/.lsr_auditd_purge_rulesd.backup", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 45, "uid": 0, "version": "3798961615", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": false, "xusr": true } } TASK [Remove rules.d tree before restoring snapshot] *************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:143 Tuesday 21 April 2026 21:26:23 -0400 (0:00:00.414) 0:00:05.856 ********* changed: [managed-node1] => { "changed": true, "path": "/etc/audit/rules.d", "state": "absent" } TASK [Recreate empty rules.d directory] **************************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:148 Tuesday 21 April 2026 21:26:24 -0400 (0:00:00.336) 0:00:06.192 ********* changed: [managed-node1] => { "changed": true, "gid": 0, "group": "root", "mode": "0750", "owner": "root", "path": "/etc/audit/rules.d", "secontext": "unconfined_u:object_r:auditd_etc_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [Copy snapshot back into rules.d] ***************************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:154 Tuesday 21 April 2026 21:26:24 -0400 (0:00:00.341) 0:00:06.533 ********* changed: [managed-node1] => { "changed": true, "checksum": null, "dest": "/etc/audit/rules.d/", "gid": 0, "group": "root", "md5sum": null, "mode": "0750", "owner": "root", "secontext": "unconfined_u:object_r:auditd_etc_t:s0", "size": 45, "src": "/root/.lsr_auditd_purge_rulesd.backup/", "state": "directory", "uid": 0 } TASK [Remove purge test rules.d snapshot directory on managed host] ************ task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:161 Tuesday 21 April 2026 21:26:24 -0400 (0:00:00.336) 0:00:06.870 ********* changed: [managed-node1] => { "changed": true, "path": "/root/.lsr_auditd_purge_rulesd.backup", "state": "absent" } TASK [Restore system state after purge test] *********************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:167 Tuesday 21 April 2026 21:26:25 -0400 (0:00:00.332) 0:00:07.202 ********* included: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/cleanup.yml for managed-node1 TASK [Restore auditd.conf from integration test backup] ************************ task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/cleanup.yml:3 Tuesday 21 April 2026 21:26:25 -0400 (0:00:00.020) 0:00:07.223 ********* ok: [managed-node1] => { "changed": false, "checksum": "33eefb6af1c751110a125b8b468118e6f0edb244", "dest": "/etc/audit/auditd.conf", "gid": 0, "group": "root", "md5sum": "eda8c45ce09c30766469af2714e2eff9", "mode": "0640", "owner": "root", "secontext": "system_u:object_r:auditd_etc_t:s0", "size": 852, "src": "/root/.lsr_auditd_integration_auditd.conf.bak", "state": "file", "uid": 0 } TASK [Remove auditd.conf backup file] ****************************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/cleanup.yml:11 Tuesday 21 April 2026 21:26:25 -0400 (0:00:00.337) 0:00:07.560 ********* changed: [managed-node1] => { "changed": true, "path": "/root/.lsr_auditd_integration_auditd.conf.bak", "state": "absent" } TASK [Restore custom.rules from integration test backup] *********************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/cleanup.yml:17 Tuesday 21 April 2026 21:26:25 -0400 (0:00:00.335) 0:00:07.896 ********* ok: [managed-node1] => { "changed": false, "checksum": "beaedc34af6f74d852b504550a00030393947a4b", "dest": "/etc/audit/rules.d/custom.rules", "gid": 0, "group": "root", "md5sum": "02a93488fb84ceabb705a049f70ced6b", "mode": "0664", "owner": "root", "secontext": "unconfined_u:object_r:auditd_etc_t:s0", "size": 293, "src": "/root/.lsr_auditd_integration_custom.rules.bak", "state": "file", "uid": 0 } TASK [Remove custom.rules backup file] ***************************************** task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/cleanup.yml:25 Tuesday 21 April 2026 21:26:26 -0400 (0:00:00.335) 0:00:08.232 ********* changed: [managed-node1] => { "changed": true, "path": "/root/.lsr_auditd_integration_custom.rules.bak", "state": "absent" } TASK [Remove custom.rules if it did not exist before test] ********************* task path: /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/cleanup.yml:31 Tuesday 21 April 2026 21:26:26 -0400 (0:00:00.335) 0:00:08.567 ********* skipping: [managed-node1] => { "changed": false, "false_condition": "not __auditd_integration_had_custom_rules | bool", "skip_reason": "Conditional result was False" } PLAY RECAP ********************************************************************* managed-node1 : ok=49 changed=11 unreachable=0 failed=1 skipped=15 rescued=0 ignored=0 SYSTEM ROLES ERRORS BEGIN v1 [ { "ansible_version": "2.16.18", "end_time": "2026-04-22T01:26:23.459204+00:00Z", "host": "managed-node1", "message": "The conditional check 'item.permission is auditd_non_empty_str_or_list' failed. The error was: template error while templating string: Could not load \"auditd_non_empty_str_or_list\": 'auditd_non_empty_str_or_list'. String: {% set __cres = item.permission is auditd_non_empty_str_or_list %}{{ [true if __cres else false, __cres.__class__.__name__] }}. Could not load \"auditd_non_empty_str_or_list\": 'auditd_non_empty_str_or_list'\n\nThe error appears to be in '/tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml': line 123, column 9, but may\nbe elsewhere in the file depending on the exact syntax problem.\n\nThe offending line appears to be:\n\n that:\n - item.permission is auditd_non_empty_str_or_list\n ^ here\n", "start_time": "2026-04-22T01:26:23.435923+00:00Z", "task_name": "Assert auditd_rules permission is a non-empty string or list of non-empty strings", "task_path": "/tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/roles/auditd/tasks/assert_auditd_rules.yml:120" } ] SYSTEM ROLES ERRORS END v1 TASKS RECAP ******************************************************************** Tuesday 21 April 2026 21:26:26 -0400 (0:00:00.012) 0:00:08.580 ********* =============================================================================== Gathering Facts --------------------------------------------------------- 1.09s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:5 Install decoy rules fragment to be removed by purge --------------------- 0.62s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:34 Back up auditd.conf before test ----------------------------------------- 0.45s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/setup.yml:21 Ensure rules.d backup parent directory is absent ------------------------ 0.44s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:16 Check for existing auditd.conf ------------------------------------------ 0.44s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/setup.yml:7 Check whether purge test rules.d snapshot exists ------------------------ 0.41s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:135 Recreate empty rules.d directory ---------------------------------------- 0.34s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:148 Back up custom.rules before test ---------------------------------------- 0.34s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/setup.yml:29 Restore auditd.conf from integration test backup ------------------------ 0.34s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/cleanup.yml:3 Copy snapshot back into rules.d ----------------------------------------- 0.34s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:154 Remove rules.d tree before restoring snapshot --------------------------- 0.34s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:143 Create parent directory for rules.d snapshot ---------------------------- 0.34s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:21 Remove custom.rules backup file ----------------------------------------- 0.34s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/cleanup.yml:25 Remove auditd.conf backup file ------------------------------------------ 0.34s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/cleanup.yml:11 Restore custom.rules from integration test backup ----------------------- 0.34s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/cleanup.yml:17 Remove purge test rules.d snapshot directory on managed host ------------ 0.33s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:161 Check for existing custom.rules ----------------------------------------- 0.33s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/setup.yml:12 Snapshot rules.d tree for full restore after test ----------------------- 0.33s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:27 Run the role normally --------------------------------------------------- 0.08s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tasks/run_role_with_clear_facts.yml:33 Back up existing auditd configuration and rules metadata ---------------- 0.07s /tmp/collections-LJi/ansible_collections/fedora/linux_system_roles/tests/auditd/tests_purge_rules.yml:13 -- Logs begin at Tue 2026-04-21 21:22:22 EDT, end at Tue 2026-04-21 21:26:26 EDT. -- Apr 21 21:26:17 managed-node1 sshd[11773]: Accepted publickey for root from 10.31.12.219 port 57626 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Apr 21 21:26:17 managed-node1 systemd[1]: Started Session 12 of user root. -- Subject: Unit session-12.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-12.scope has finished starting up. -- -- The start-up result is done. Apr 21 21:26:17 managed-node1 systemd-logind[603]: New session 12 of user root. -- Subject: A new session 12 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 12 has been created for the user root. -- -- The leading process of the session is 11773. Apr 21 21:26:17 managed-node1 sshd[11773]: pam_unix(sshd:session): session opened for user root by (uid=0) Apr 21 21:26:17 managed-node1 sshd[11776]: Received disconnect from 10.31.12.219 port 57626:11: disconnected by user Apr 21 21:26:17 managed-node1 sshd[11776]: Disconnected from user root 10.31.12.219 port 57626 Apr 21 21:26:17 managed-node1 sshd[11773]: pam_unix(sshd:session): session closed for user root Apr 21 21:26:17 managed-node1 systemd[1]: session-12.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit session-12.scope has successfully entered the 'dead' state. Apr 21 21:26:17 managed-node1 systemd-logind[603]: Session 12 logged out. Waiting for processes to exit. Apr 21 21:26:17 managed-node1 systemd-logind[603]: Removed session 12. -- Subject: Session 12 has been terminated -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 12 has been terminated. Apr 21 21:26:18 managed-node1 platform-python[11938]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Apr 21 21:26:19 managed-node1 platform-python[12090]: ansible-ansible.builtin.stat Invoked with path=/etc/audit/auditd.conf follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Apr 21 21:26:19 managed-node1 platform-python[12215]: ansible-ansible.builtin.stat Invoked with path=/etc/audit/rules.d/custom.rules follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Apr 21 21:26:20 managed-node1 platform-python[12340]: ansible-ansible.legacy.copy Invoked with src=/etc/audit/auditd.conf dest=/root/.lsr_auditd_integration_auditd.conf.bak remote_src=True mode=preserve backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:20 managed-node1 platform-python[12465]: ansible-ansible.legacy.copy Invoked with src=/etc/audit/rules.d/custom.rules dest=/root/.lsr_auditd_integration_custom.rules.bak remote_src=True mode=preserve backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:21 managed-node1 platform-python[12590]: ansible-ansible.builtin.file Invoked with path=/root/.lsr_auditd_purge_rulesd.backup state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:21 managed-node1 platform-python[12713]: ansible-ansible.builtin.file Invoked with path=/root/.lsr_auditd_purge_rulesd.backup state=directory mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:21 managed-node1 platform-python[12836]: ansible-ansible.legacy.copy Invoked with src=/etc/audit/rules.d/ dest=/root/.lsr_auditd_purge_rulesd.backup/ remote_src=True mode=preserve backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:22 managed-node1 platform-python[12959]: ansible-ansible.legacy.stat Invoked with path=/etc/audit/rules.d/99-lsr-purge-decoy.rules follow=False get_checksum=True checksum_algorithm=sha1 get_mime=True get_attributes=True Apr 21 21:26:22 managed-node1 platform-python[13058]: ansible-ansible.legacy.copy Invoked with dest=/etc/audit/rules.d/99-lsr-purge-decoy.rules mode=0644 src=/root/.ansible/tmp/ansible-tmp-1776821181.9469764-9982-220291465627703/source _original_basename=tmp3_zr8usd follow=False checksum=35324fa08aa315fc1ab79570382f077fec44dce9 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:23 managed-node1 platform-python[13222]: ansible-ansible.builtin.stat Invoked with path=/root/.lsr_auditd_purge_rulesd.backup follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Apr 21 21:26:24 managed-node1 platform-python[13347]: ansible-ansible.builtin.file Invoked with path=/etc/audit/rules.d state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:24 managed-node1 platform-python[13470]: ansible-ansible.builtin.file Invoked with path=/etc/audit/rules.d state=directory mode=0750 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:24 managed-node1 platform-python[13593]: ansible-ansible.legacy.copy Invoked with src=/root/.lsr_auditd_purge_rulesd.backup/ dest=/etc/audit/rules.d/ remote_src=True mode=preserve backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:25 managed-node1 platform-python[13716]: ansible-ansible.builtin.file Invoked with path=/root/.lsr_auditd_purge_rulesd.backup state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:25 managed-node1 platform-python[13839]: ansible-ansible.legacy.copy Invoked with src=/root/.lsr_auditd_integration_auditd.conf.bak dest=/etc/audit/auditd.conf remote_src=True mode=preserve backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:25 managed-node1 platform-python[13962]: ansible-ansible.builtin.file Invoked with path=/root/.lsr_auditd_integration_auditd.conf.bak state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:26 managed-node1 platform-python[14085]: ansible-ansible.legacy.copy Invoked with src=/root/.lsr_auditd_integration_custom.rules.bak dest=/etc/audit/rules.d/custom.rules remote_src=True mode=preserve backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:26 managed-node1 platform-python[14208]: ansible-ansible.builtin.file Invoked with path=/root/.lsr_auditd_integration_custom.rules.bak state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Apr 21 21:26:26 managed-node1 sshd[14229]: Accepted publickey for root from 10.31.12.219 port 43670 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Apr 21 21:26:26 managed-node1 systemd[1]: Started Session 13 of user root. -- Subject: Unit session-13.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-13.scope has finished starting up. -- -- The start-up result is done. Apr 21 21:26:26 managed-node1 systemd-logind[603]: New session 13 of user root. -- Subject: A new session 13 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 13 has been created for the user root. -- -- The leading process of the session is 14229. Apr 21 21:26:26 managed-node1 sshd[14229]: pam_unix(sshd:session): session opened for user root by (uid=0) Apr 21 21:26:26 managed-node1 sshd[14232]: Received disconnect from 10.31.12.219 port 43670:11: disconnected by user Apr 21 21:26:26 managed-node1 sshd[14232]: Disconnected from user root 10.31.12.219 port 43670 Apr 21 21:26:26 managed-node1 sshd[14229]: pam_unix(sshd:session): session closed for user root Apr 21 21:26:26 managed-node1 systemd[1]: session-13.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit session-13.scope has successfully entered the 'dead' state. Apr 21 21:26:26 managed-node1 systemd-logind[603]: Session 13 logged out. Waiting for processes to exit. Apr 21 21:26:26 managed-node1 systemd-logind[603]: Removed session 13. -- Subject: Session 13 has been terminated -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 13 has been terminated. Apr 21 21:26:26 managed-node1 sshd[14253]: Accepted publickey for root from 10.31.12.219 port 43678 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Apr 21 21:26:26 managed-node1 systemd[1]: Started Session 14 of user root. -- Subject: Unit session-14.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-14.scope has finished starting up. -- -- The start-up result is done. Apr 21 21:26:26 managed-node1 systemd-logind[603]: New session 14 of user root. -- Subject: A new session 14 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 14 has been created for the user root. -- -- The leading process of the session is 14253. Apr 21 21:26:26 managed-node1 sshd[14253]: pam_unix(sshd:session): session opened for user root by (uid=0)