[WARNING]: Collection infra.leapp does not support Ansible version 2.14.18 [WARNING]: running playbook inside collection infra.leapp [WARNING]: Collection community.general does not support Ansible version 2.14.18 ansible-playbook [core 2.14.18] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /root/.ansible/collections:/usr/share/ansible/collections executable location = /usr/bin/ansible-playbook python version = 3.9.23 (main, Aug 19 2025, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-11)] (/usr/bin/python3) jinja version = 3.1.2 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_default.yml **************************************************** 1 plays in /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/tests_default.yml PLAY [Test] ******************************************************************** TASK [Gathering Facts] ********************************************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/tests_default.yml:2 ok: [managed-node01] TASK [Initialize lock, logging, and common vars] ******************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:5 TASK [infra.leapp.common : Log directory exists] ******************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:3 changed: [managed-node01] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/var/log/ripu", "secontext": "unconfined_u:object_r:var_log_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.common : Check for existing log file] ************************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:11 ok: [managed-node01] => {"changed": false, "stat": {"exists": false}} TASK [infra.leapp.common : Fail if log file already exists] ******************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:16 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : Create new log file] ******************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:23 NOTIFIED HANDLER infra.leapp.common : Add end time to log file for managed-node01 NOTIFIED HANDLER infra.leapp.common : Slurp ripu.log file for managed-node01 NOTIFIED HANDLER infra.leapp.common : Decode ripu.log file for managed-node01 NOTIFIED HANDLER infra.leapp.common : Rename log file for managed-node01 changed: [managed-node01] => {"changed": true, "checksum": "a332a9047b8a2654bda5fc96a97b09e490da807e", "dest": "/var/log/ripu/ripu.log", "gid": 0, "group": "root", "md5sum": "7da973ebb435a49d7774b1a34e2967bc", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:var_log_t:s0", "size": 61, "src": "/root/.ansible/tmp/ansible-tmp-1762892688.6796744-7641-274602471951693/source", "state": "file", "uid": 0} TASK [infra.leapp.common : /etc/ansible/facts.d directory exists] ************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:35 changed: [managed-node01] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/ansible/facts.d", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.common : Capture current ansible_facts for validation after upgrade] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:43 changed: [managed-node01] => {"changed": true, "checksum": "f3c145f024e19eaea2640ab8297d2bf09b15d512", "dest": "/etc/ansible/facts.d/pre_ripu.fact", "gid": 0, "group": "root", "md5sum": "8672c49d74301a0024c4448eb38fb8f0", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 13830, "src": "/root/.ansible/tmp/ansible-tmp-1762892689.8233943-7669-132438657874190/source", "state": "file", "uid": 0} TASK [infra.leapp.common : Capture a list of non-rhel versioned packages] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:51 ok: [managed-node01] => {"changed": false, "cmd": "set -o pipefail; export PATH=$PATH; rpm -qa | grep -ve '[\\.|+]el8' | grep -vE '^(gpg-pubkey|libmodulemd|katello-ca-consumer)' | sort", "delta": "0:00:00.877850", "end": "2025-11-11 15:24:51.737186", "failed_when_result": false, "msg": "non-zero return code", "rc": 1, "start": "2025-11-11 15:24:50.859336", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [infra.leapp.common : Create fact with the non-rhel versioned packages list] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:65 ok: [managed-node01] => {"ansible_facts": {"non_rhel_packages": []}, "changed": false} TASK [infra.leapp.common : Capture the list of non-rhel versioned packages in a separate fact file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:69 changed: [managed-node01] => {"changed": true, "checksum": "97d170e1550eee4afc0af065b78cda302a97674c", "dest": "/etc/ansible/facts.d/non_rhel_packages.fact", "gid": 0, "group": "root", "md5sum": "d751713988987e9331980363e24189ce", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 2, "src": "/root/.ansible/tmp/ansible-tmp-1762892691.8554535-7702-93048619686807/source", "state": "file", "uid": 0} TASK [infra.leapp.analysis : Include tasks for preupg assistant analysis] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:12 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : Include tasks for leapp preupgrade analysis] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:16 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml for managed-node01 TASK [infra.leapp.analysis : analysis-leapp | Register to leapp activation key] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:2 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [analysis-leapp | Include custom_local_repos for local_repos_pre_leapp] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:14 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Install packages for preupgrade analysis on RHEL 7] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:22 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Install packages for preupgrade analysis on RHEL 8] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:29 changed: [managed-node01] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: systemd-container-239-82.el8_10.8.x86_64", "Installed: python3-leapp-0.19.0-1.el8_10.noarch", "Installed: leapp-0.19.0-1.el8_10.noarch", "Installed: leapp-deps-0.19.0-1.el8_10.noarch", "Installed: leapp-upgrade-el8toel9-0.22.0-1.el8_10.noarch", "Installed: leapp-upgrade-el8toel9-deps-0.22.0-1.el8_10.noarch", "Installed: policycoreutils-python-utils-2.9-26.el8_10.noarch"]} TASK [infra.leapp.analysis : analysis-leapp | Install packages for preupgrade analysis on RHEL 9] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:36 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Ensure leapp log directory exists] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:43 ok: [managed-node01] => {"changed": false, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/var/log/leapp", "secontext": "system_u:object_r:var_log_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.analysis : analysis-leapp | Populate leapp_answers file] ***** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:51 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [analysis-leapp | Create /etc/leapp/files/leapp_upgrade_repositories.repo] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:60 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Leapp preupgrade report] ********* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:71 ASYNC FAILED on managed-node01: jid=j35206699581.7829 changed: [managed-node01] => {"ansible_job_id": "j35206699581.7829", "changed": true, "cmd": "set -o pipefail; export PATH=$PATH; ulimit -n 16384; leapp preupgrade --report-schema=1.2.0 2>&1 | tee -a /var/log/ripu/ripu.log\n", "delta": "0:00:40.447099", "end": "2025-11-11 15:25:45.385285", "failed_when_result": false, "finished": 1, "msg": "non-zero return code", "rc": 1, "results_file": "/root/.ansible_async/j35206699581.7829", "start": "2025-11-11 15:25:04.938186", "started": 1, "stderr": "", "stderr_lines": [], "stdout": "==> Processing phase `configuration_phase`\n====> * ipu_workflow_config\n IPU workflow config actor\n==> Processing phase `FactsCollection`\n====> * scan_kernel_cmdline\n No documentation has been provided for the scan_kernel_cmdline actor.\n====> * scanmemory\n Scan Memory of the machine.\n====> * scan_source_files\n Scan files (explicitly specified) of the source system.\n====> * copy_dnf_conf_into_target_userspace\n Copy dnf.conf into target userspace\n====> * repository_mapping\n Produces message containing repository mapping based on provided file.\n====> * storage_scanner\n Provides data about storage settings.\n====> * scan_custom_modifications_actor\n Collects information about files in leapp directories that have been modified or newly added.\n====> * scancryptopolicies\n Scan information about system wide set crypto policies including:\n====> * rpm_scanner\n Provides data about installed RPM Packages.\n====> * system_facts\n Provides data about many facts from system.\n====> * trusted_gpg_keys_scanner\n Scan for trusted GPG keys.\n====> * sssd_facts_8to9\n Check SSSD configuration for changes in RHEL9 and report them in model.\n====> * xfs_info_scanner\n This actor scans all mounted mountpoints for XFS information.\n====> * read_openssh_config\n Collect information about the OpenSSH configuration.\n====> * transaction_workarounds\n Provides additional RPM transaction tasks based on bundled RPM packages.\n====> * luks_scanner\n Provides data about active LUKS devices.\n====> * scan_pkg_manager\n Provides data about package manager (yum/dnf)\n====> * ifcfg_scanner\n Scan ifcfg files with legacy network configuration\n====> * nis_scanner\n Collect information about the NIS packages configuration.\n====> * scandasd\n In case of s390x architecture, check whether DASD is used.\n====> * scan_target_os_image\n Scans the provided target OS ISO image to use as a content source for the IPU, if any.\n====> * scan_grub_config\n Scan grub configuration files for errors.\n====> * firewalld_collect_used_object_names\n This actor reads firewalld's configuration and produces Model\n====> * get_enabled_modules\n Provides data about which module streams are enabled on the source system.\n====> * check_custom_network_scripts\n Check the existence of custom network-scripts and warn user about possible\n====> * scanzfcp\n In case of s390x architecture, check whether ZFCP is used.\n====> * network_manager_connection_scanner\n Scan NetworkManager connection keyfiles\n====> * scan_custom_repofile\n Scan the custom /etc/leapp/files/leapp_upgrade_repositories.repo repo file.\n====> * roce_scanner\n Detect active RoCE NICs on IBM Z machines.\n====> * scanclienablerepo\n Produce CustomTargetRepository based on the LEAPP_ENABLE_REPOS in config.\n====> * remove_obsolete_gpg_keys\n Remove obsoleted RPM GPG keys.\n====> * scan_defined_ipu_paths\n Load defined IPU paths for the current major source system version\n====> * scanblacklistca\n Scan the file system for distrusted CA's in the blacklist directory.\n====> * scan_sap_hana\n Gathers information related to SAP HANA instances on the system.\n====> * scan_systemd_source\n Provides info about systemd on the source system\n====> * scan_grub_device_name\n Find the name of the block devices where GRUB is located\n====> * distribution_signed_rpm_scanner\n Provide data about distribution signed & unsigned RPM packages.\n====> * open_ssl_config_scanner\n Read an OpenSSL configuration file for further analysis.\n====> * xorgdrvfacts8to9\n Check the journal logs for deprecated Xorg drivers.\n====> * checkrhui\n Check if system is using RHUI infrastructure (on public cloud) and send messages to\n====> * root_scanner\n Scan the system root directory and produce a message containing\n====> * persistentnetnames\n Get network interface information for physical ethernet interfaces of the original system.\n====> * register_ruby_irb_adjustment\n Register a workaround to allow rubygem-irb's directory -> symlink conversion.\n====> * scan_files_for_target_userspace\n Scan the source system and identify files that will be copied into the target userspace when it is created.\n====> * ipa_scanner\n Scan system for ipa-client and ipa-server status\n====> * load_device_driver_deprecation_data\n Loads deprecation data for drivers and devices (PCI & CPU)\n====> * network_manager_read_config\n Provides data about NetworkManager configuration.\n====> * scan_subscription_manager_info\n Scans the current system for subscription manager information\n====> * get_installed_desktops\n Actor checks if kde or gnome desktop environments\n====> * udevadm_info\n Produces data exported by the \"udevadm info\" command.\n====> * selinuxcontentscanner\n Scan the system for any SELinux customizations\n====> * firewalld_collect_global_config\n This actor reads firewalld's configuration and produces Model\n====> * scan_fips\n Determine whether the source system has FIPS enabled.\n====> * used_repository_scanner\n Scan used enabled repositories\n====> * scan_source_kernel\n Scan the source system kernel.\n====> * detect_kernel_drivers\n Matches all currently loaded kernel drivers against known deprecated and removed drivers.\n====> * pci_devices_scanner\n Provides data about existing PCI Devices.\n====> * scancpu\n Scan CPUs of the machine.\n====> * rpm_transaction_config_tasks_collector\n Provides additional RPM transaction tasks from /etc/leapp/transaction.\n====> * repositories_blacklist\n Exclude target repositories provided by Red Hat without support.\n====> * biosdevname\n Enable biosdevname on the target RHEL system if all interfaces on the source RHEL\n====> * scan_dynamic_linker_configuration\n Scan the dynamic linker configuration and find modifications.\n====> * multipath_conf_read_8to9\n Read multipath configuration files and extract the necessary information\n====> * persistentnetnamesdisable\n Disable systemd-udevd persistent network naming on machine with single eth0 NIC\n====> * vdo_conversion_scanner\n Provides conversion info about VDO devices.\n====> * satellite_upgrade_facts\n Report which Satellite packages require updates and how to handle PostgreSQL data\n====> * satellite_upgrade_services\n Reconfigure Satellite services\n====> * pes_events_scanner\n Provides data about package events from Package Evolution Service.\n====> * setuptargetrepos\n Produces list of repositories that should be available to be used by Upgrade process.\n\n============================================================\n ERRORS \n============================================================\n\n2025-11-11 15:25:30.375635 [ERROR] Actor: scan_subscription_manager_info\nMessage: A subscription-manager command failed to execute\nSummary:\n Details: Command ['subscription-manager', 'release'] failed with exit code 1.\n Stderr: This system is not yet registered. Try 'subscription-manager register --help' for more information.\n Hint: Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\n Link: https://access.redhat.com/solutions/6138372\n\n============================================================\n END OF ERRORS \n============================================================\n\nDebug output written to /var/log/leapp/leapp-preupgrade.log\n\n============================================================\n REPORT OVERVIEW \n============================================================\n\nFollowing errors occurred and the upgrade cannot continue:\n 1. Actor: scan_subscription_manager_info\n Message: A subscription-manager command failed to execute\n\nHIGH and MEDIUM severity reports:\n 1. Packages available in excluded repositories will not be installed\n\nReports summary:\n Errors: 1\n Inhibitors: 0\n HIGH severity reports: 1\n MEDIUM severity reports: 0\n LOW severity reports: 0\n INFO severity reports: 1\n\nBefore continuing, review the full report below for details about discovered problems and possible remediation instructions:\n A report has been generated at /var/log/leapp/leapp-report.txt\n A report has been generated at /var/log/leapp/leapp-report.json\n\n============================================================\n END OF REPORT OVERVIEW \n============================================================\n\nAnswerfile has been generated at /var/log/leapp/answerfile", "stdout_lines": ["==> Processing phase `configuration_phase`", "====> * ipu_workflow_config", " IPU workflow config actor", "==> Processing phase `FactsCollection`", "====> * scan_kernel_cmdline", " No documentation has been provided for the scan_kernel_cmdline actor.", "====> * scanmemory", " Scan Memory of the machine.", "====> * scan_source_files", " Scan files (explicitly specified) of the source system.", "====> * copy_dnf_conf_into_target_userspace", " Copy dnf.conf into target userspace", "====> * repository_mapping", " Produces message containing repository mapping based on provided file.", "====> * storage_scanner", " Provides data about storage settings.", "====> * scan_custom_modifications_actor", " Collects information about files in leapp directories that have been modified or newly added.", "====> * scancryptopolicies", " Scan information about system wide set crypto policies including:", "====> * rpm_scanner", " Provides data about installed RPM Packages.", "====> * system_facts", " Provides data about many facts from system.", "====> * trusted_gpg_keys_scanner", " Scan for trusted GPG keys.", "====> * sssd_facts_8to9", " Check SSSD configuration for changes in RHEL9 and report them in model.", "====> * xfs_info_scanner", " This actor scans all mounted mountpoints for XFS information.", "====> * read_openssh_config", " Collect information about the OpenSSH configuration.", "====> * transaction_workarounds", " Provides additional RPM transaction tasks based on bundled RPM packages.", "====> * luks_scanner", " Provides data about active LUKS devices.", "====> * scan_pkg_manager", " Provides data about package manager (yum/dnf)", "====> * ifcfg_scanner", " Scan ifcfg files with legacy network configuration", "====> * nis_scanner", " Collect information about the NIS packages configuration.", "====> * scandasd", " In case of s390x architecture, check whether DASD is used.", "====> * scan_target_os_image", " Scans the provided target OS ISO image to use as a content source for the IPU, if any.", "====> * scan_grub_config", " Scan grub configuration files for errors.", "====> * firewalld_collect_used_object_names", " This actor reads firewalld's configuration and produces Model", "====> * get_enabled_modules", " Provides data about which module streams are enabled on the source system.", "====> * check_custom_network_scripts", " Check the existence of custom network-scripts and warn user about possible", "====> * scanzfcp", " In case of s390x architecture, check whether ZFCP is used.", "====> * network_manager_connection_scanner", " Scan NetworkManager connection keyfiles", "====> * scan_custom_repofile", " Scan the custom /etc/leapp/files/leapp_upgrade_repositories.repo repo file.", "====> * roce_scanner", " Detect active RoCE NICs on IBM Z machines.", "====> * scanclienablerepo", " Produce CustomTargetRepository based on the LEAPP_ENABLE_REPOS in config.", "====> * remove_obsolete_gpg_keys", " Remove obsoleted RPM GPG keys.", "====> * scan_defined_ipu_paths", " Load defined IPU paths for the current major source system version", "====> * scanblacklistca", " Scan the file system for distrusted CA's in the blacklist directory.", "====> * scan_sap_hana", " Gathers information related to SAP HANA instances on the system.", "====> * scan_systemd_source", " Provides info about systemd on the source system", "====> * scan_grub_device_name", " Find the name of the block devices where GRUB is located", "====> * distribution_signed_rpm_scanner", " Provide data about distribution signed & unsigned RPM packages.", "====> * open_ssl_config_scanner", " Read an OpenSSL configuration file for further analysis.", "====> * xorgdrvfacts8to9", " Check the journal logs for deprecated Xorg drivers.", "====> * checkrhui", " Check if system is using RHUI infrastructure (on public cloud) and send messages to", "====> * root_scanner", " Scan the system root directory and produce a message containing", "====> * persistentnetnames", " Get network interface information for physical ethernet interfaces of the original system.", "====> * register_ruby_irb_adjustment", " Register a workaround to allow rubygem-irb's directory -> symlink conversion.", "====> * scan_files_for_target_userspace", " Scan the source system and identify files that will be copied into the target userspace when it is created.", "====> * ipa_scanner", " Scan system for ipa-client and ipa-server status", "====> * load_device_driver_deprecation_data", " Loads deprecation data for drivers and devices (PCI & CPU)", "====> * network_manager_read_config", " Provides data about NetworkManager configuration.", "====> * scan_subscription_manager_info", " Scans the current system for subscription manager information", "====> * get_installed_desktops", " Actor checks if kde or gnome desktop environments", "====> * udevadm_info", " Produces data exported by the \"udevadm info\" command.", "====> * selinuxcontentscanner", " Scan the system for any SELinux customizations", "====> * firewalld_collect_global_config", " This actor reads firewalld's configuration and produces Model", "====> * scan_fips", " Determine whether the source system has FIPS enabled.", "====> * used_repository_scanner", " Scan used enabled repositories", "====> * scan_source_kernel", " Scan the source system kernel.", "====> * detect_kernel_drivers", " Matches all currently loaded kernel drivers against known deprecated and removed drivers.", "====> * pci_devices_scanner", " Provides data about existing PCI Devices.", "====> * scancpu", " Scan CPUs of the machine.", "====> * rpm_transaction_config_tasks_collector", " Provides additional RPM transaction tasks from /etc/leapp/transaction.", "====> * repositories_blacklist", " Exclude target repositories provided by Red Hat without support.", "====> * biosdevname", " Enable biosdevname on the target RHEL system if all interfaces on the source RHEL", "====> * scan_dynamic_linker_configuration", " Scan the dynamic linker configuration and find modifications.", "====> * multipath_conf_read_8to9", " Read multipath configuration files and extract the necessary information", "====> * persistentnetnamesdisable", " Disable systemd-udevd persistent network naming on machine with single eth0 NIC", "====> * vdo_conversion_scanner", " Provides conversion info about VDO devices.", "====> * satellite_upgrade_facts", " Report which Satellite packages require updates and how to handle PostgreSQL data", "====> * satellite_upgrade_services", " Reconfigure Satellite services", "====> * pes_events_scanner", " Provides data about package events from Package Evolution Service.", "====> * setuptargetrepos", " Produces list of repositories that should be available to be used by Upgrade process.", "", "============================================================", " ERRORS ", "============================================================", "", "2025-11-11 15:25:30.375635 [ERROR] Actor: scan_subscription_manager_info", "Message: A subscription-manager command failed to execute", "Summary:", " Details: Command ['subscription-manager', 'release'] failed with exit code 1.", " Stderr: This system is not yet registered. Try 'subscription-manager register --help' for more information.", " Hint: Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.", " Link: https://access.redhat.com/solutions/6138372", "", "============================================================", " END OF ERRORS ", "============================================================", "", "Debug output written to /var/log/leapp/leapp-preupgrade.log", "", "============================================================", " REPORT OVERVIEW ", "============================================================", "", "Following errors occurred and the upgrade cannot continue:", " 1. Actor: scan_subscription_manager_info", " Message: A subscription-manager command failed to execute", "", "HIGH and MEDIUM severity reports:", " 1. Packages available in excluded repositories will not be installed", "", "Reports summary:", " Errors: 1", " Inhibitors: 0", " HIGH severity reports: 1", " MEDIUM severity reports: 0", " LOW severity reports: 0", " INFO severity reports: 1", "", "Before continuing, review the full report below for details about discovered problems and possible remediation instructions:", " A report has been generated at /var/log/leapp/leapp-report.txt", " A report has been generated at /var/log/leapp/leapp-report.json", "", "============================================================", " END OF REPORT OVERVIEW ", "============================================================", "", "Answerfile has been generated at /var/log/leapp/answerfile"]} TASK [analysis-leapp | Include custom_local_repos for local_repos_post_analysis] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:89 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Include check-results-file.yml] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:97 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/check-results-file.yml for managed-node01 TASK [infra.leapp.analysis : check-results-file | Result file status] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/check-results-file.yml:2 ok: [managed-node01] => {"changed": false, "stat": {"atime": 1762892744.6971552, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "72ca5190ff1a2541590a7c12d80beccf2f61cf53", "ctime": 1762892744.6981554, "dev": 51715, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 360710313, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1762892744.6981554, "nlink": 1, "path": "/var/log/leapp/leapp-report.txt", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 2696, "uid": 0, "version": "2729802387", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}} TASK [infra.leapp.analysis : check-results-file | Check that result file exists] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/check-results-file.yml:7 ok: [managed-node01] => { "changed": false, "msg": "All assertions passed" } TASK [analysis-leapp | Include the parse_leapp_report role to check for inhibitors] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:100 TASK [infra.leapp.parse_leapp_report : Default upgrade_inhibited to false] ***** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/parse_leapp_report/tasks/main.yml:2 ok: [managed-node01] => {"ansible_facts": {"upgrade_inhibited": false}, "changed": false} TASK [infra.leapp.parse_leapp_report : Collect human readable report results] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/parse_leapp_report/tasks/main.yml:6 ok: [managed-node01] => {"changed": false, "content": "UmlzayBGYWN0b3I6IGhpZ2ggKGVycm9yKQpUaXRsZTogQSBzdWJzY3JpcHRpb24tbWFuYWdlciBjb21tYW5kIGZhaWxlZCB0byBleGVjdXRlClN1bW1hcnk6IHsiZGV0YWlscyI6ICJDb21tYW5kIFsnc3Vic2NyaXB0aW9uLW1hbmFnZXInLCAncmVsZWFzZSddIGZhaWxlZCB3aXRoIGV4aXQgY29kZSAxLiIsICJzdGRlcnIiOiAiVGhpcyBzeXN0ZW0gaXMgbm90IHlldCByZWdpc3RlcmVkLiBUcnkgJ3N1YnNjcmlwdGlvbi1tYW5hZ2VyIHJlZ2lzdGVyIC0taGVscCcgZm9yIG1vcmUgaW5mb3JtYXRpb24uXG4iLCAiaGludCI6ICJQbGVhc2UgZW5zdXJlIHlvdSBoYXZlIGEgdmFsaWQgUkhFTCBzdWJzY3JpcHRpb24gYW5kIHlvdXIgbmV0d29yayBpcyB1cC4gSWYgeW91IGFyZSB1c2luZyBwcm94eSBmb3IgUmVkIEhhdCBzdWJzY3JpcHRpb24tbWFuYWdlciwgcGxlYXNlIG1ha2Ugc3VyZSBpdCBpcyBzcGVjaWZpZWQgaW5zaWRlIHRoZSAvZXRjL3Joc20vcmhzbS5jb25mIGZpbGUuIE9yIHVzZSB0aGUgLS1uby1yaHNtIG9wdGlvbiB3aGVuIHJ1bm5pbmcgbGVhcHAsIGlmIHlvdSBkbyBub3Qgd2FudCB0byB1c2Ugc3Vic2NyaXB0aW9uLW1hbmFnZXIgZm9yIHRoZSBpbi1wbGFjZSB1cGdyYWRlIGFuZCB5b3Ugd2FudCB0byBkZWxpdmVyIGFsbCB0YXJnZXQgcmVwb3NpdG9yaWVzIGJ5IHlvdXJzZWxmIG9yIHVzaW5nIFJIVUkgb24gcHVibGljIGNsb3VkLiIsICJsaW5rIjogImh0dHBzOi8vYWNjZXNzLnJlZGhhdC5jb20vc29sdXRpb25zLzYxMzgzNzIifQpLZXk6IDdlYzgyNjk3ODRkYjFiYmEyYWM1NGFlNDM4Njg5ZWYzOTdlMTY4MzMKLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLQpSaXNrIEZhY3RvcjogaGlnaCAKVGl0bGU6IFBhY2thZ2VzIGF2YWlsYWJsZSBpbiBleGNsdWRlZCByZXBvc2l0b3JpZXMgd2lsbCBub3QgYmUgaW5zdGFsbGVkClN1bW1hcnk6IDEgcGFja2FnZXMgd2lsbCBiZSBza2lwcGVkIGJlY2F1c2UgdGhleSBhcmUgYXZhaWxhYmxlIG9ubHkgaW4gdGFyZ2V0IHN5c3RlbSByZXBvc2l0b3JpZXMgdGhhdCBhcmUgaW50ZW50aW9uYWxseSBleGNsdWRlZCBmcm9tIHRoZSBsaXN0IG9mIHJlcG9zaXRvcmllcyB1c2VkIGR1cmluZyB0aGUgdXBncmFkZS4gU2VlIHRoZSByZXBvcnQgbWVzc2FnZSB0aXRsZWQgIkV4Y2x1ZGVkIHRhcmdldCBzeXN0ZW0gcmVwb3NpdG9yaWVzIiBmb3IgZGV0YWlscy4KVGhlIGxpc3Qgb2YgdGhlc2UgcGFja2FnZXM6Ci0gaml0dGVyZW50cm9weS1kZXZlbCAocmVwb2lkOiBjb2RlcmVhZHktYnVpbGRlci1mb3ItcmhlbC05LXg4Nl82NC1ycG1zKQpLZXk6IDI0MzdlMjA0ODA4Zjk4NzQ3N2MwZTliZThlNGM5NWIzYTg3YTlmM2UKLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLQpSaXNrIEZhY3RvcjogaW5mbyAKVGl0bGU6IEV4Y2x1ZGVkIHRhcmdldCBzeXN0ZW0gcmVwb3NpdG9yaWVzClN1bW1hcnk6IFRoZSBmb2xsb3dpbmcgcmVwb3NpdG9yaWVzIGFyZSBub3Qgc3VwcG9ydGVkIGJ5IFJlZCBIYXQgYW5kIGFyZSBleGNsdWRlZCBmcm9tIHRoZSBsaXN0IG9mIHJlcG9zaXRvcmllcyB1c2VkIGR1cmluZyB0aGUgdXBncmFkZS4KLSBjb2RlcmVhZHktYnVpbGRlci1mb3ItcmhlbC05LWFhcmNoNjQtcnBtcwotIHJodWktY29kZXJlYWR5LWJ1aWxkZXItZm9yLXJoZWwtOS1hYXJjaDY0LXJodWktcnBtcwotIGNvZGVyZWFkeS1idWlsZGVyLWZvci1yaGVsLTktcmh1aS1ycG1zCi0gY29kZXJlYWR5LWJ1aWxkZXItZm9yLXJoZWwtOS1wcGM2NGxlLWV1cy1ycG1zCi0gY29kZXJlYWR5LWJ1aWxkZXItZm9yLXJoZWwtOS14ODZfNjQtcmh1aS1ycG1zCi0gcmh1aS1jb2RlcmVhZHktYnVpbGRlci1mb3ItcmhlbC05LXg4Nl82NC1yaHVpLXJwbXMKLSBjb2RlcmVhZHktYnVpbGRlci1mb3ItcmhlbC05LXMzOTB4LXJwbXMKLSBjb2RlcmVhZHktYnVpbGRlci1mb3ItcmhlbC05LXMzOTB4LWV1cy1ycG1zCi0gY29kZXJlYWR5LWJ1aWxkZXItYmV0YS1mb3ItcmhlbC05LXg4Nl82NC1ycG1zCi0gY29kZXJlYWR5LWJ1aWxkZXItZm9yLXJoZWwtOS14ODZfNjQtZXVzLXJwbXMKLSBjb2RlcmVhZHktYnVpbGRlci1mb3ItcmhlbC05LWFhcmNoNjQtZXVzLXJwbXMKLSBjb2RlcmVhZHktYnVpbGRlci1iZXRhLWZvci1yaGVsLTktcHBjNjRsZS1ycG1zCi0gY29kZXJlYWR5LWJ1aWxkZXItZm9yLXJoZWwtOS14ODZfNjQtcnBtcwotIGNvZGVyZWFkeS1idWlsZGVyLWZvci1yaGVsLTktcHBjNjRsZS1ycG1zCi0gY29kZXJlYWR5LWJ1aWxkZXItYmV0YS1mb3ItcmhlbC05LWFhcmNoNjQtcnBtcwotIGNvZGVyZWFkeS1idWlsZGVyLWJldGEtZm9yLXJoZWwtOS1zMzkweC1ycG1zClJlbWVkaWF0aW9uOiBbaGludF0gSWYgc29tZSBvZiBleGNsdWRlZCByZXBvc2l0b3JpZXMgYXJlIHN0aWxsIHJlcXVpcmVkIHRvIGJlIHVzZWQgZHVyaW5nIHRoZSB1cGdyYWRlLCBleGVjdXRlIGxlYXBwIHdpdGggdGhlIC0tZW5hYmxlcmVwbyBvcHRpb24gd2l0aCB0aGUgcmVwb2lkIG9mIHRoZSByZXBvc2l0b3J5IHJlcXVpcmVkIHRvIGJlIGVuYWJsZWQgYXMgYW4gYXJndW1lbnQgKHRoZSBvcHRpb24gY2FuIGJlIHVzZWQgbXVsdGlwbGUgdGltZXMpLgpLZXk6IDFiOTEzMmNiMjM2MmFlNzgzMGU0OGVlZTc4MTFiZTk1Mjc3NDdkZTgKLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLS0tLQo=", "encoding": "base64", "source": "/var/log/leapp/leapp-report.txt"} TASK [infra.leapp.parse_leapp_report : Collect JSON report results] ************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/parse_leapp_report/tasks/main.yml:11 ok: [managed-node01] => {"changed": false, "content": "ewogICJlbnRyaWVzIjogWwogICAgewogICAgICAiYXVkaWVuY2UiOiAic3lzYWRtaW4iLAogICAgICAiZ3JvdXBzIjogWwogICAgICAgICJlcnJvciIKICAgICAgXSwKICAgICAgImtleSI6ICI3ZWM4MjY5Nzg0ZGIxYmJhMmFjNTRhZTQzODY4OWVmMzk3ZTE2ODMzIiwKICAgICAgInNldmVyaXR5IjogImhpZ2giLAogICAgICAic3VtbWFyeSI6ICJ7XCJkZXRhaWxzXCI6IFwiQ29tbWFuZCBbJ3N1YnNjcmlwdGlvbi1tYW5hZ2VyJywgJ3JlbGVhc2UnXSBmYWlsZWQgd2l0aCBleGl0IGNvZGUgMS5cIiwgXCJzdGRlcnJcIjogXCJUaGlzIHN5c3RlbSBpcyBub3QgeWV0IHJlZ2lzdGVyZWQuIFRyeSAnc3Vic2NyaXB0aW9uLW1hbmFnZXIgcmVnaXN0ZXIgLS1oZWxwJyBmb3IgbW9yZSBpbmZvcm1hdGlvbi5cXG5cIiwgXCJoaW50XCI6IFwiUGxlYXNlIGVuc3VyZSB5b3UgaGF2ZSBhIHZhbGlkIFJIRUwgc3Vic2NyaXB0aW9uIGFuZCB5b3VyIG5ldHdvcmsgaXMgdXAuIElmIHlvdSBhcmUgdXNpbmcgcHJveHkgZm9yIFJlZCBIYXQgc3Vic2NyaXB0aW9uLW1hbmFnZXIsIHBsZWFzZSBtYWtlIHN1cmUgaXQgaXMgc3BlY2lmaWVkIGluc2lkZSB0aGUgL2V0Yy9yaHNtL3Joc20uY29uZiBmaWxlLiBPciB1c2UgdGhlIC0tbm8tcmhzbSBvcHRpb24gd2hlbiBydW5uaW5nIGxlYXBwLCBpZiB5b3UgZG8gbm90IHdhbnQgdG8gdXNlIHN1YnNjcmlwdGlvbi1tYW5hZ2VyIGZvciB0aGUgaW4tcGxhY2UgdXBncmFkZSBhbmQgeW91IHdhbnQgdG8gZGVsaXZlciBhbGwgdGFyZ2V0IHJlcG9zaXRvcmllcyBieSB5b3Vyc2VsZiBvciB1c2luZyBSSFVJIG9uIHB1YmxpYyBjbG91ZC5cIiwgXCJsaW5rXCI6IFwiaHR0cHM6Ly9hY2Nlc3MucmVkaGF0LmNvbS9zb2x1dGlvbnMvNjEzODM3MlwifSIsCiAgICAgICJ0aXRsZSI6ICJBIHN1YnNjcmlwdGlvbi1tYW5hZ2VyIGNvbW1hbmQgZmFpbGVkIHRvIGV4ZWN1dGUiLAogICAgICAidGltZVN0YW1wIjogIjIwMjUtMTEtMTFUMjA6MjU6MzAuMzc1ODkyWiIsCiAgICAgICJob3N0bmFtZSI6ICJtYW5hZ2VkLW5vZGUwMSIsCiAgICAgICJhY3RvciI6ICJzY2FuX3N1YnNjcmlwdGlvbl9tYW5hZ2VyX2luZm8iLAogICAgICAiaWQiOiAiODE0OGM3ZmVjZjMzYjBmMDg4YWU5OGVhMWJmNzc1ZjQ4NTAzYTQxODJjNWQ1Mjk0YjIxODU4Y2E5MmJiYTQ1NyIKICAgIH0sCiAgICB7CiAgICAgICJhdWRpZW5jZSI6ICJzeXNhZG1pbiIsCiAgICAgICJkZXRhaWwiOiB7CiAgICAgICAgInJlbWVkaWF0aW9ucyI6IFsKICAgICAgICAgIHsKICAgICAgICAgICAgImNvbnRleHQiOiAiSWYgc29tZSBvZiBleGNsdWRlZCByZXBvc2l0b3JpZXMgYXJlIHN0aWxsIHJlcXVpcmVkIHRvIGJlIHVzZWQgZHVyaW5nIHRoZSB1cGdyYWRlLCBleGVjdXRlIGxlYXBwIHdpdGggdGhlIC0tZW5hYmxlcmVwbyBvcHRpb24gd2l0aCB0aGUgcmVwb2lkIG9mIHRoZSByZXBvc2l0b3J5IHJlcXVpcmVkIHRvIGJlIGVuYWJsZWQgYXMgYW4gYXJndW1lbnQgKHRoZSBvcHRpb24gY2FuIGJlIHVzZWQgbXVsdGlwbGUgdGltZXMpLiIsCiAgICAgICAgICAgICJ0eXBlIjogImhpbnQiCiAgICAgICAgICB9CiAgICAgICAgXQogICAgICB9LAogICAgICAiZ3JvdXBzIjogWwogICAgICAgICJyZXBvc2l0b3J5IiwKICAgICAgICAiZmFpbHVyZSIKICAgICAgXSwKICAgICAgImtleSI6ICIxYjkxMzJjYjIzNjJhZTc4MzBlNDhlZWU3ODExYmU5NTI3NzQ3ZGU4IiwKICAgICAgInNldmVyaXR5IjogImluZm8iLAogICAgICAic3VtbWFyeSI6ICJUaGUgZm9sbG93aW5nIHJlcG9zaXRvcmllcyBhcmUgbm90IHN1cHBvcnRlZCBieSBSZWQgSGF0IGFuZCBhcmUgZXhjbHVkZWQgZnJvbSB0aGUgbGlzdCBvZiByZXBvc2l0b3JpZXMgdXNlZCBkdXJpbmcgdGhlIHVwZ3JhZGUuXG4tIGNvZGVyZWFkeS1idWlsZGVyLWZvci1yaGVsLTktYWFyY2g2NC1ycG1zXG4tIHJodWktY29kZXJlYWR5LWJ1aWxkZXItZm9yLXJoZWwtOS1hYXJjaDY0LXJodWktcnBtc1xuLSBjb2RlcmVhZHktYnVpbGRlci1mb3ItcmhlbC05LXJodWktcnBtc1xuLSBjb2RlcmVhZHktYnVpbGRlci1mb3ItcmhlbC05LXBwYzY0bGUtZXVzLXJwbXNcbi0gY29kZXJlYWR5LWJ1aWxkZXItZm9yLXJoZWwtOS14ODZfNjQtcmh1aS1ycG1zXG4tIHJodWktY29kZXJlYWR5LWJ1aWxkZXItZm9yLXJoZWwtOS14ODZfNjQtcmh1aS1ycG1zXG4tIGNvZGVyZWFkeS1idWlsZGVyLWZvci1yaGVsLTktczM5MHgtcnBtc1xuLSBjb2RlcmVhZHktYnVpbGRlci1mb3ItcmhlbC05LXMzOTB4LWV1cy1ycG1zXG4tIGNvZGVyZWFkeS1idWlsZGVyLWJldGEtZm9yLXJoZWwtOS14ODZfNjQtcnBtc1xuLSBjb2RlcmVhZHktYnVpbGRlci1mb3ItcmhlbC05LXg4Nl82NC1ldXMtcnBtc1xuLSBjb2RlcmVhZHktYnVpbGRlci1mb3ItcmhlbC05LWFhcmNoNjQtZXVzLXJwbXNcbi0gY29kZXJlYWR5LWJ1aWxkZXItYmV0YS1mb3ItcmhlbC05LXBwYzY0bGUtcnBtc1xuLSBjb2RlcmVhZHktYnVpbGRlci1mb3ItcmhlbC05LXg4Nl82NC1ycG1zXG4tIGNvZGVyZWFkeS1idWlsZGVyLWZvci1yaGVsLTktcHBjNjRsZS1ycG1zXG4tIGNvZGVyZWFkeS1idWlsZGVyLWJldGEtZm9yLXJoZWwtOS1hYXJjaDY0LXJwbXNcbi0gY29kZXJlYWR5LWJ1aWxkZXItYmV0YS1mb3ItcmhlbC05LXMzOTB4LXJwbXMiLAogICAgICAidGl0bGUiOiAiRXhjbHVkZWQgdGFyZ2V0IHN5c3RlbSByZXBvc2l0b3JpZXMiLAogICAgICAidGltZVN0YW1wIjogIjIwMjUtMTEtMTFUMjA6MjU6NDIuNzczMDAzWiIsCiAgICAgICJob3N0bmFtZSI6ICJtYW5hZ2VkLW5vZGUwMSIsCiAgICAgICJhY3RvciI6ICJyZXBvc2l0b3JpZXNfYmxhY2tsaXN0IiwKICAgICAgImlkIjogImFhODhhODQzOTg4YTY5Nzc1OTM4MTllMzc0ZDU1YzIwNGIzOGE0YzNiNGExNmYwODEzNzQyNGE0YzgxYzc1MDciCiAgICB9LAogICAgewogICAgICAiYXVkaWVuY2UiOiAic3lzYWRtaW4iLAogICAgICAiZGV0YWlsIjogewogICAgICAgICJyZWxhdGVkX3Jlc291cmNlcyI6IFsKICAgICAgICAgIHsKICAgICAgICAgICAgInNjaGVtZSI6ICJwYWNrYWdlIiwKICAgICAgICAgICAgInRpdGxlIjogImppdHRlcmVudHJvcHktZGV2ZWwiCiAgICAgICAgICB9CiAgICAgICAgXQogICAgICB9LAogICAgICAiZ3JvdXBzIjogWwogICAgICAgICJyZXBvc2l0b3J5IgogICAgICBdLAogICAgICAia2V5IjogIjI0MzdlMjA0ODA4Zjk4NzQ3N2MwZTliZThlNGM5NWIzYTg3YTlmM2UiLAogICAgICAic2V2ZXJpdHkiOiAiaGlnaCIsCiAgICAgICJzdW1tYXJ5IjogIjEgcGFja2FnZXMgd2lsbCBiZSBza2lwcGVkIGJlY2F1c2UgdGhleSBhcmUgYXZhaWxhYmxlIG9ubHkgaW4gdGFyZ2V0IHN5c3RlbSByZXBvc2l0b3JpZXMgdGhhdCBhcmUgaW50ZW50aW9uYWxseSBleGNsdWRlZCBmcm9tIHRoZSBsaXN0IG9mIHJlcG9zaXRvcmllcyB1c2VkIGR1cmluZyB0aGUgdXBncmFkZS4gU2VlIHRoZSByZXBvcnQgbWVzc2FnZSB0aXRsZWQgXCJFeGNsdWRlZCB0YXJnZXQgc3lzdGVtIHJlcG9zaXRvcmllc1wiIGZvciBkZXRhaWxzLlxuVGhlIGxpc3Qgb2YgdGhlc2UgcGFja2FnZXM6XG4tIGppdHRlcmVudHJvcHktZGV2ZWwgKHJlcG9pZDogY29kZXJlYWR5LWJ1aWxkZXItZm9yLXJoZWwtOS14ODZfNjQtcnBtcykiLAogICAgICAidGl0bGUiOiAiUGFja2FnZXMgYXZhaWxhYmxlIGluIGV4Y2x1ZGVkIHJlcG9zaXRvcmllcyB3aWxsIG5vdCBiZSBpbnN0YWxsZWQiLAogICAgICAidGltZVN0YW1wIjogIjIwMjUtMTEtMTFUMjA6MjU6NDQuNDkwNzcxWiIsCiAgICAgICJob3N0bmFtZSI6ICJtYW5hZ2VkLW5vZGUwMSIsCiAgICAgICJhY3RvciI6ICJwZXNfZXZlbnRzX3NjYW5uZXIiLAogICAgICAiaWQiOiAiZWNhZjllOTlkMDIxZTY0NTMyYTA1MTY2NDU1NjY5NDFmZGU5MGFmODA5NjQyMjRmMTI0M2UwZDA3MmE5ZTQ0OSIKICAgIH0KICBdLAogICJsZWFwcF9ydW5faWQiOiAiY2ZjNjQzYmMtMzU3Mi00NDNjLTlmMzktMzNmMGFhNjBmNzQyIgp9Cg==", "encoding": "base64", "source": "/var/log/leapp/leapp-report.json"} TASK [infra.leapp.parse_leapp_report : Parse report results] ******************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/parse_leapp_report/tasks/main.yml:16 ok: [managed-node01] => {"ansible_facts": {"leapp_report_json": {"entries": [{"actor": "scan_subscription_manager_info", "audience": "sysadmin", "groups": ["error"], "hostname": "managed-node01", "id": "8148c7fecf33b0f088ae98ea1bf775f48503a4182c5d5294b21858ca92bba457", "key": "7ec8269784db1bba2ac54ae438689ef397e16833", "severity": "high", "summary": "{\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "timeStamp": "2025-11-11T20:25:30.375892Z", "title": "A subscription-manager command failed to execute"}, {"actor": "repositories_blacklist", "audience": "sysadmin", "detail": {"remediations": [{"context": "If some of excluded repositories are still required to be used during the upgrade, execute leapp with the --enablerepo option with the repoid of the repository required to be enabled as an argument (the option can be used multiple times).", "type": "hint"}]}, "groups": ["repository", "failure"], "hostname": "managed-node01", "id": "aa88a843988a6977593819e374d55c204b38a4c3b4a16f08137424a4c81c7507", "key": "1b9132cb2362ae7830e48eee7811be9527747de8", "severity": "info", "summary": "The following repositories are not supported by Red Hat and are excluded from the list of repositories used during the upgrade.\n- codeready-builder-for-rhel-9-aarch64-rpms\n- rhui-codeready-builder-for-rhel-9-aarch64-rhui-rpms\n- codeready-builder-for-rhel-9-rhui-rpms\n- codeready-builder-for-rhel-9-ppc64le-eus-rpms\n- codeready-builder-for-rhel-9-x86_64-rhui-rpms\n- rhui-codeready-builder-for-rhel-9-x86_64-rhui-rpms\n- codeready-builder-for-rhel-9-s390x-rpms\n- codeready-builder-for-rhel-9-s390x-eus-rpms\n- codeready-builder-beta-for-rhel-9-x86_64-rpms\n- codeready-builder-for-rhel-9-x86_64-eus-rpms\n- codeready-builder-for-rhel-9-aarch64-eus-rpms\n- codeready-builder-beta-for-rhel-9-ppc64le-rpms\n- codeready-builder-for-rhel-9-x86_64-rpms\n- codeready-builder-for-rhel-9-ppc64le-rpms\n- codeready-builder-beta-for-rhel-9-aarch64-rpms\n- codeready-builder-beta-for-rhel-9-s390x-rpms", "timeStamp": "2025-11-11T20:25:42.773003Z", "title": "Excluded target system repositories"}, {"actor": "pes_events_scanner", "audience": "sysadmin", "detail": {"related_resources": [{"scheme": "package", "title": "jitterentropy-devel"}]}, "groups": ["repository"], "hostname": "managed-node01", "id": "ecaf9e99d021e64532a0516645566941fde90af80964224f1243e0d072a9e449", "key": "2437e204808f987477c0e9be8e4c95b3a87a9f3e", "severity": "high", "summary": "1 packages will be skipped because they are available only in target system repositories that are intentionally excluded from the list of repositories used during the upgrade. See the report message titled \"Excluded target system repositories\" for details.\nThe list of these packages:\n- jitterentropy-devel (repoid: codeready-builder-for-rhel-9-x86_64-rpms)", "timeStamp": "2025-11-11T20:25:44.490771Z", "title": "Packages available in excluded repositories will not be installed"}], "leapp_run_id": "cfc643bc-3572-443c-9f39-33f0aa60f742"}, "leapp_report_txt": ["Risk Factor: high (error)", "Title: A subscription-manager command failed to execute", "Summary: {\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "Key: 7ec8269784db1bba2ac54ae438689ef397e16833", "----------------------------------------", "Risk Factor: high ", "Title: Packages available in excluded repositories will not be installed", "Summary: 1 packages will be skipped because they are available only in target system repositories that are intentionally excluded from the list of repositories used during the upgrade. See the report message titled \"Excluded target system repositories\" for details.", "The list of these packages:", "- jitterentropy-devel (repoid: codeready-builder-for-rhel-9-x86_64-rpms)", "Key: 2437e204808f987477c0e9be8e4c95b3a87a9f3e", "----------------------------------------", "Risk Factor: info ", "Title: Excluded target system repositories", "Summary: The following repositories are not supported by Red Hat and are excluded from the list of repositories used during the upgrade.", "- codeready-builder-for-rhel-9-aarch64-rpms", "- rhui-codeready-builder-for-rhel-9-aarch64-rhui-rpms", "- codeready-builder-for-rhel-9-rhui-rpms", "- codeready-builder-for-rhel-9-ppc64le-eus-rpms", "- codeready-builder-for-rhel-9-x86_64-rhui-rpms", "- rhui-codeready-builder-for-rhel-9-x86_64-rhui-rpms", "- codeready-builder-for-rhel-9-s390x-rpms", "- codeready-builder-for-rhel-9-s390x-eus-rpms", "- codeready-builder-beta-for-rhel-9-x86_64-rpms", "- codeready-builder-for-rhel-9-x86_64-eus-rpms", "- codeready-builder-for-rhel-9-aarch64-eus-rpms", "- codeready-builder-beta-for-rhel-9-ppc64le-rpms", "- codeready-builder-for-rhel-9-x86_64-rpms", "- codeready-builder-for-rhel-9-ppc64le-rpms", "- codeready-builder-beta-for-rhel-9-aarch64-rpms", "- codeready-builder-beta-for-rhel-9-s390x-rpms", "Remediation: [hint] If some of excluded repositories are still required to be used during the upgrade, execute leapp with the --enablerepo option with the repoid of the repository required to be enabled as an argument (the option can be used multiple times).", "Key: 1b9132cb2362ae7830e48eee7811be9527747de8", "----------------------------------------", ""]}, "changed": false} TASK [infra.leapp.parse_leapp_report : Check for inhibitors] ******************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/parse_leapp_report/tasks/main.yml:22 ok: [managed-node01] => (item={'audience': 'sysadmin', 'groups': ['error'], 'key': '7ec8269784db1bba2ac54ae438689ef397e16833', 'severity': 'high', 'summary': '{"details": "Command [\'subscription-manager\', \'release\'] failed with exit code 1.", "stderr": "This system is not yet registered. Try \'subscription-manager register --help\' for more information.\\n", "hint": "Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.", "link": "https://access.redhat.com/solutions/6138372"}', 'title': 'A subscription-manager command failed to execute', 'timeStamp': '2025-11-11T20:25:30.375892Z', 'hostname': 'managed-node01', 'actor': 'scan_subscription_manager_info', 'id': '8148c7fecf33b0f088ae98ea1bf775f48503a4182c5d5294b21858ca92bba457'}) => {"ansible_facts": {"leapp_inhibitors": [{"actor": "scan_subscription_manager_info", "audience": "sysadmin", "groups": ["error"], "hostname": "managed-node01", "id": "8148c7fecf33b0f088ae98ea1bf775f48503a4182c5d5294b21858ca92bba457", "key": "7ec8269784db1bba2ac54ae438689ef397e16833", "severity": "high", "summary": "{\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "timeStamp": "2025-11-11T20:25:30.375892Z", "title": "A subscription-manager command failed to execute"}], "upgrade_inhibited": true}, "ansible_loop_var": "item", "changed": false, "item": {"actor": "scan_subscription_manager_info", "audience": "sysadmin", "groups": ["error"], "hostname": "managed-node01", "id": "8148c7fecf33b0f088ae98ea1bf775f48503a4182c5d5294b21858ca92bba457", "key": "7ec8269784db1bba2ac54ae438689ef397e16833", "severity": "high", "summary": "{\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "timeStamp": "2025-11-11T20:25:30.375892Z", "title": "A subscription-manager command failed to execute"}} skipping: [managed-node01] => (item={'audience': 'sysadmin', 'detail': {'remediations': [{'context': 'If some of excluded repositories are still required to be used during the upgrade, execute leapp with the --enablerepo option with the repoid of the repository required to be enabled as an argument (the option can be used multiple times).', 'type': 'hint'}]}, 'groups': ['repository', 'failure'], 'key': '1b9132cb2362ae7830e48eee7811be9527747de8', 'severity': 'info', 'summary': 'The following repositories are not supported by Red Hat and are excluded from the list of repositories used during the upgrade.\n- codeready-builder-for-rhel-9-aarch64-rpms\n- rhui-codeready-builder-for-rhel-9-aarch64-rhui-rpms\n- codeready-builder-for-rhel-9-rhui-rpms\n- codeready-builder-for-rhel-9-ppc64le-eus-rpms\n- codeready-builder-for-rhel-9-x86_64-rhui-rpms\n- rhui-codeready-builder-for-rhel-9-x86_64-rhui-rpms\n- codeready-builder-for-rhel-9-s390x-rpms\n- codeready-builder-for-rhel-9-s390x-eus-rpms\n- codeready-builder-beta-for-rhel-9-x86_64-rpms\n- codeready-builder-for-rhel-9-x86_64-eus-rpms\n- codeready-builder-for-rhel-9-aarch64-eus-rpms\n- codeready-builder-beta-for-rhel-9-ppc64le-rpms\n- codeready-builder-for-rhel-9-x86_64-rpms\n- codeready-builder-for-rhel-9-ppc64le-rpms\n- codeready-builder-beta-for-rhel-9-aarch64-rpms\n- codeready-builder-beta-for-rhel-9-s390x-rpms', 'title': 'Excluded target system repositories', 'timeStamp': '2025-11-11T20:25:42.773003Z', 'hostname': 'managed-node01', 'actor': 'repositories_blacklist', 'id': 'aa88a843988a6977593819e374d55c204b38a4c3b4a16f08137424a4c81c7507'}) => {"ansible_loop_var": "item", "changed": false, "item": {"actor": "repositories_blacklist", "audience": "sysadmin", "detail": {"remediations": [{"context": "If some of excluded repositories are still required to be used during the upgrade, execute leapp with the --enablerepo option with the repoid of the repository required to be enabled as an argument (the option can be used multiple times).", "type": "hint"}]}, "groups": ["repository", "failure"], "hostname": "managed-node01", "id": "aa88a843988a6977593819e374d55c204b38a4c3b4a16f08137424a4c81c7507", "key": "1b9132cb2362ae7830e48eee7811be9527747de8", "severity": "info", "summary": "The following repositories are not supported by Red Hat and are excluded from the list of repositories used during the upgrade.\n- codeready-builder-for-rhel-9-aarch64-rpms\n- rhui-codeready-builder-for-rhel-9-aarch64-rhui-rpms\n- codeready-builder-for-rhel-9-rhui-rpms\n- codeready-builder-for-rhel-9-ppc64le-eus-rpms\n- codeready-builder-for-rhel-9-x86_64-rhui-rpms\n- rhui-codeready-builder-for-rhel-9-x86_64-rhui-rpms\n- codeready-builder-for-rhel-9-s390x-rpms\n- codeready-builder-for-rhel-9-s390x-eus-rpms\n- codeready-builder-beta-for-rhel-9-x86_64-rpms\n- codeready-builder-for-rhel-9-x86_64-eus-rpms\n- codeready-builder-for-rhel-9-aarch64-eus-rpms\n- codeready-builder-beta-for-rhel-9-ppc64le-rpms\n- codeready-builder-for-rhel-9-x86_64-rpms\n- codeready-builder-for-rhel-9-ppc64le-rpms\n- codeready-builder-beta-for-rhel-9-aarch64-rpms\n- codeready-builder-beta-for-rhel-9-s390x-rpms", "timeStamp": "2025-11-11T20:25:42.773003Z", "title": "Excluded target system repositories"}, "skip_reason": "Conditional result was False"} skipping: [managed-node01] => (item={'audience': 'sysadmin', 'detail': {'related_resources': [{'scheme': 'package', 'title': 'jitterentropy-devel'}]}, 'groups': ['repository'], 'key': '2437e204808f987477c0e9be8e4c95b3a87a9f3e', 'severity': 'high', 'summary': '1 packages will be skipped because they are available only in target system repositories that are intentionally excluded from the list of repositories used during the upgrade. See the report message titled "Excluded target system repositories" for details.\nThe list of these packages:\n- jitterentropy-devel (repoid: codeready-builder-for-rhel-9-x86_64-rpms)', 'title': 'Packages available in excluded repositories will not be installed', 'timeStamp': '2025-11-11T20:25:44.490771Z', 'hostname': 'managed-node01', 'actor': 'pes_events_scanner', 'id': 'ecaf9e99d021e64532a0516645566941fde90af80964224f1243e0d072a9e449'}) => {"ansible_loop_var": "item", "changed": false, "item": {"actor": "pes_events_scanner", "audience": "sysadmin", "detail": {"related_resources": [{"scheme": "package", "title": "jitterentropy-devel"}]}, "groups": ["repository"], "hostname": "managed-node01", "id": "ecaf9e99d021e64532a0516645566941fde90af80964224f1243e0d072a9e449", "key": "2437e204808f987477c0e9be8e4c95b3a87a9f3e", "severity": "high", "summary": "1 packages will be skipped because they are available only in target system repositories that are intentionally excluded from the list of repositories used during the upgrade. See the report message titled \"Excluded target system repositories\" for details.\nThe list of these packages:\n- jitterentropy-devel (repoid: codeready-builder-for-rhel-9-x86_64-rpms)", "timeStamp": "2025-11-11T20:25:44.490771Z", "title": "Packages available in excluded repositories will not be installed"}, "skip_reason": "Conditional result was False"} TASK [infra.leapp.parse_leapp_report : Collect inhibitors] ********************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/parse_leapp_report/tasks/main.yml:34 ok: [managed-node01] => {"changed": false, "cmd": ["awk", "/\\(inhibitor\\)/,/^-------/", "/var/log/leapp/leapp-report.txt"], "delta": "0:00:00.004114", "end": "2025-11-11 15:26:07.357408", "failed_when_result": false, "msg": "", "rc": 0, "start": "2025-11-11 15:26:07.353294", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [infra.leapp.parse_leapp_report : Collect high errors] ******************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/parse_leapp_report/tasks/main.yml:43 ok: [managed-node01] => {"changed": false, "cmd": ["awk", "/high \\(error\\)/,/^-------/", "/var/log/leapp/leapp-report.txt"], "delta": "0:00:00.004444", "end": "2025-11-11 15:26:07.701360", "failed_when_result": false, "msg": "", "rc": 0, "start": "2025-11-11 15:26:07.696916", "stderr": "", "stderr_lines": [], "stdout": "Risk Factor: high (error)\nTitle: A subscription-manager command failed to execute\nSummary: {\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}\nKey: 7ec8269784db1bba2ac54ae438689ef397e16833\n----------------------------------------", "stdout_lines": ["Risk Factor: high (error)", "Title: A subscription-manager command failed to execute", "Summary: {\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "Key: 7ec8269784db1bba2ac54ae438689ef397e16833", "----------------------------------------"]} TASK [infra.leapp.analysis : Set stats for leapp_inhibitors] ******************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:20 ok: [managed-node01] => {"ansible_stats": {"aggregate": true, "data": {"leapp_inhibitors": [{"actor": "scan_subscription_manager_info", "audience": "sysadmin", "groups": ["error"], "hostname": "managed-node01", "id": "8148c7fecf33b0f088ae98ea1bf775f48503a4182c5d5294b21858ca92bba457", "key": "7ec8269784db1bba2ac54ae438689ef397e16833", "severity": "high", "summary": "{\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "timeStamp": "2025-11-11T20:25:30.375892Z", "title": "A subscription-manager command failed to execute"}]}, "per_host": false}, "changed": false} TASK [infra.leapp.analysis : Notify analysis report is done handler] *********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:25 NOTIFIED HANDLER infra.leapp.analysis : Preupgrade analysis report is done for managed-node01 NOTIFIED HANDLER infra.leapp.analysis : Display inhibitors for managed-node01 NOTIFIED HANDLER infra.leapp.analysis : Display errors for managed-node01 changed: [managed-node01] => {"changed": true, "msg": "All assertions passed"} RUNNING HANDLER [infra.leapp.analysis : Display inhibitors] ******************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/handlers/main.yml:32 skipping: [managed-node01] => {} RUNNING HANDLER [infra.leapp.analysis : Display errors] ************************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/handlers/main.yml:40 ok: [managed-node01] => { "results_errors.stdout_lines": [ "Risk Factor: high (error)", "Title: A subscription-manager command failed to execute", "Summary: {\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "Key: 7ec8269784db1bba2ac54ae438689ef397e16833", "----------------------------------------" ] } RUNNING HANDLER [infra.leapp.analysis : Preupgrade analysis report is done] **** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/handlers/main.yml:48 ok: [managed-node01] => { "msg": "The preupgrade analysis report generation is now complete. WARNING: Inhibitors found. Review the tasks above or the result file at /var/log/leapp/leapp-report.txt." } RUNNING HANDLER [infra.leapp.common : Add end time to log file] **************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/handlers/main.yml:11 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} RUNNING HANDLER [infra.leapp.common : Slurp ripu.log file] ********************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/handlers/main.yml:21 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} RUNNING HANDLER [infra.leapp.common : Decode ripu.log file] ******************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/handlers/main.yml:28 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} RUNNING HANDLER [infra.leapp.common : Rename log file] ************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/handlers/main.yml:34 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} PLAY RECAP ********************************************************************* managed-node01 : ok=27 changed=8 unreachable=0 failed=0 skipped=14 rescued=0 ignored=0