[WARNING]: Collection infra.leapp does not support Ansible version 2.14.18 [WARNING]: running playbook inside collection infra.leapp [WARNING]: Collection community.general does not support Ansible version 2.14.18 ansible-playbook [core 2.14.18] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /root/.ansible/collections:/usr/share/ansible/collections executable location = /usr/bin/ansible-playbook python version = 3.9.23 (main, Aug 19 2025, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-11)] (/usr/bin/python3) jinja version = 3.1.2 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_default.yml **************************************************** 1 plays in /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/tests_default.yml PLAY [Test] ******************************************************************** TASK [Gathering Facts] ********************************************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/tests_default.yml:2 ok: [managed-node01] TASK [Initialize lock, logging, and common vars] ******************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:5 TASK [infra.leapp.common : Log directory exists] ******************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:3 ok: [managed-node01] => {"changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/var/log/ripu", "secontext": "unconfined_u:object_r:var_log_t:s0", "size": 22, "state": "directory", "uid": 0} TASK [infra.leapp.common : Check for existing log file] ************************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:11 ok: [managed-node01] => {"changed": false, "stat": {"atime": 1762892539.9440694, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "dce5a3f0fd71f6511c4ad4ad403dbdabb5cc52f4", "ctime": 1762892539.9470694, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 58720344, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1762892539.4490614, "nlink": 1, "path": "/var/log/ripu/ripu.log", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 61, "uid": 0, "version": "18446744072669229232", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}} TASK [infra.leapp.common : Fail if log file already exists] ******************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:16 fatal: [managed-node01]: FAILED! => {"changed": false, "msg": "Another RIPU playbook job is already running. See /var/log/ripu/ripu.log for details. If the previous job was aborted, rename the log file to clear this failure and try again."} PLAY RECAP ********************************************************************* managed-node01 : ok=3 changed=0 unreachable=0 failed=1 skipped=0 rescued=0 ignored=0 -- Logs begin at Tue 2025-11-11 15:18:33 EST, end at Tue 2025-11-11 15:23:22 EST. -- Nov 11 15:23:20 managed-node01 ansible-ansible.legacy.setup[4852]: Invoked with filter=[] gather_subset=['all'] fact_path=/etc/ansible/facts.d gather_timeout=10 Nov 11 15:23:21 managed-node01 ansible-ansible.builtin.file[4942]: Invoked with src=None selevel=None force=False setype=None _original_basename=None unsafe_writes=False access_time=None seuser=None recurse=False state=directory access_time_format=%Y%m%d%H%M.%S group=root modification_time=None serole=None _diff_peek=None modification_time_format=%Y%m%d%H%M.%S path=/var/log/ripu owner=root follow=True attributes=None mode=0755 Nov 11 15:23:22 managed-node01 ansible-ansible.builtin.stat[5003]: Invoked with checksum_algorithm=sha1 get_checksum=True follow=False path=/var/log/ripu/ripu.log get_md5=False get_mime=True get_attributes=True Nov 11 15:23:22 managed-node01 sshd[5016]: Accepted publickey for root from 10.31.46.171 port 48006 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Nov 11 15:23:22 managed-node01 systemd-logind[531]: New session 9 of user root. -- Subject: A new session 9 has been created for user root -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- Documentation: http://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 9 has been created for the user root. -- -- The leading process of the session is 5016. Nov 11 15:23:22 managed-node01 systemd[1]: Started Session 9 of user root. -- Subject: Unit session-9.scope has finished start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit session-9.scope has finished starting up. -- -- The start-up result is done. Nov 11 15:23:22 managed-node01 sshd[5016]: pam_unix(sshd:session): session opened for user root by (uid=0) Nov 11 15:23:22 managed-node01 sshd[5016]: Received disconnect from 10.31.46.171 port 48006:11: disconnected by user Nov 11 15:23:22 managed-node01 sshd[5016]: Disconnected from 10.31.46.171 port 48006 Nov 11 15:23:22 managed-node01 sshd[5016]: pam_unix(sshd:session): session closed for user root Nov 11 15:23:22 managed-node01 systemd-logind[531]: Removed session 9. -- Subject: Session 9 has been terminated -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- Documentation: http://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 9 has been terminated. Nov 11 15:23:22 managed-node01 sshd[5028]: Accepted publickey for root from 10.31.46.171 port 43830 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Nov 11 15:23:22 managed-node01 systemd[1]: Started Session 10 of user root. -- Subject: Unit session-10.scope has finished start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit session-10.scope has finished starting up. -- -- The start-up result is done. Nov 11 15:23:22 managed-node01 systemd-logind[531]: New session 10 of user root. -- Subject: A new session 10 has been created for user root -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- Documentation: http://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 10 has been created for the user root. -- -- The leading process of the session is 5028. Nov 11 15:23:22 managed-node01 sshd[5028]: pam_unix(sshd:session): session opened for user root by (uid=0)