ansible-playbook [core 2.16.14]
config file = None
configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.12/site-packages/ansible
ansible collection location = /tmp/collections-SQL
executable location = /usr/local/bin/ansible-playbook
python version = 3.12.1 (main, Feb 21 2024, 14:18:26) [GCC 8.5.0 20210514 (Red Hat 8.5.0-21)] (/usr/bin/python3.12)
jinja version = 3.1.4
libyaml = True
No config file found; using defaults
running playbook inside collection fedora.linux_system_roles
redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug
redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug
redirecting (type: callback) ansible.builtin.profile_tasks to ansible.posix.profile_tasks
Skipping callback 'default', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.
PLAYBOOK: tests_cib_properties_empty.yml ***************************************
2 plays in /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cib_properties_empty.yml
PLAY [all] *********************************************************************
TASK [Include vault variables] *************************************************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cib_properties_empty.yml:5
Wednesday 11 December 2024 11:27:17 -0500 (0:00:00.007) 0:00:00.007 ****
ok: [managed-node1] => {
"ansible_facts": {
"ha_cluster_hacluster_password": {
"__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n31303833633366333561656439323930303361333161363239346166656537323933313436\n3432386236656563343237306335323637396239616230353561330a313731623238393238\n62343064666336643930663239383936616465643134646536656532323461356237646133\n3761616633323839633232353637366266350a313163633236376666653238633435306565\n3264623032333736393535663833\n"
}
},
"ansible_included_var_files": [
"/tmp/ha_cluster-iZO/tests/vars/vault-variables.yml"
],
"changed": false
}
PLAY [Configure cluster properties - empty set] ********************************
TASK [Gathering Facts] *********************************************************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cib_properties_empty.yml:9
Wednesday 11 December 2024 11:27:17 -0500 (0:00:00.017) 0:00:00.024 ****
ok: [managed-node1]
TASK [Set up test environment] *************************************************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cib_properties_empty.yml:23
Wednesday 11 December 2024 11:27:18 -0500 (0:00:01.152) 0:00:01.177 ****
TASK [fedora.linux_system_roles.ha_cluster : Set node name to 'localhost' for single-node clusters] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:9
Wednesday 11 December 2024 11:27:18 -0500 (0:00:00.048) 0:00:01.225 ****
ok: [managed-node1] => {
"ansible_facts": {
"inventory_hostname": "localhost"
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Ensure facts used by tests] *******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:14
Wednesday 11 December 2024 11:27:18 -0500 (0:00:00.036) 0:00:01.261 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "'distribution' not in ansible_facts",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Check if system is ostree] ********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:22
Wednesday 11 December 2024 11:27:18 -0500 (0:00:00.015) 0:00:01.277 ****
ok: [managed-node1] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.ha_cluster : Set flag to indicate system is ostree] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:27
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.458) 0:00:01.736 ****
ok: [managed-node1] => {
"ansible_facts": {
"__ha_cluster_is_ostree": false
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Do not try to enable RHEL repositories] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:32
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.022) 0:00:01.758 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ansible_distribution == 'RedHat'",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Copy nss-altfiles ha_cluster users to /etc/passwd] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:41
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.013) 0:00:01.772 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__ha_cluster_is_ostree | d(false)",
"skip_reason": "Conditional result was False"
}
TASK [Run HA Cluster role] *****************************************************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cib_properties_empty.yml:28
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.035) 0:00:01.808 ****
TASK [fedora.linux_system_roles.ha_cluster : Set platform/version specific variables] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:3
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.050) 0:00:01.858 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Ensure ansible_facts used by role] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:2
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.023) 0:00:01.881 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__ha_cluster_required_facts | difference(ansible_facts.keys() | list) | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Check if system is ostree] ********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:10
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.035) 0:00:01.917 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "not __ha_cluster_is_ostree is defined",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Set flag to indicate system is ostree] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:15
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.019) 0:00:01.937 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "not __ha_cluster_is_ostree is defined",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Set platform/version specific variables] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:19
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.018) 0:00:01.955 ****
ok: [managed-node1] => (item=RedHat.yml) => {
"ansible_facts": {
"__ha_cluster_cloud_agents_packages": [],
"__ha_cluster_fence_agent_packages_default": "{{ ['fence-agents-all'] + (['fence-virt'] if ansible_architecture == 'x86_64' else []) }}",
"__ha_cluster_fullstack_node_packages": [
"corosync",
"libknet1-plugins-all",
"resource-agents",
"pacemaker"
],
"__ha_cluster_pcs_provider": "pcs-0.10",
"__ha_cluster_qdevice_node_packages": [
"corosync-qdevice",
"bash",
"coreutils",
"curl",
"grep",
"nss-tools",
"openssl",
"sed"
],
"__ha_cluster_repos": [],
"__ha_cluster_role_essential_packages": [
"pcs",
"corosync-qnetd",
"openssl"
],
"__ha_cluster_sbd_packages": [
"sbd"
],
"__ha_cluster_services": [
"corosync",
"corosync-qdevice",
"pacemaker"
]
},
"ansible_included_var_files": [
"/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/RedHat.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "RedHat.yml"
}
skipping: [managed-node1] => (item=CentOS.yml) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__vars_file is file",
"item": "CentOS.yml",
"skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
"ansible_facts": {
"__ha_cluster_cloud_agents_packages": [
"resource-agents-aliyun",
"resource-agents-gcp",
"fence-agents-aliyun",
"fence-agents-aws",
"fence-agents-azure-arm",
"fence-agents-gce"
],
"__ha_cluster_repos": [
{
"id": "ha",
"name": "HighAvailability"
},
{
"id": "resilientstorage",
"name": "ResilientStorage"
}
]
},
"ansible_included_var_files": [
"/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/CentOS_8.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
"ansible_facts": {
"__ha_cluster_cloud_agents_packages": [
"resource-agents-aliyun",
"resource-agents-gcp",
"fence-agents-aliyun",
"fence-agents-aws",
"fence-agents-azure-arm",
"fence-agents-gce"
],
"__ha_cluster_repos": [
{
"id": "ha",
"name": "HighAvailability"
},
{
"id": "resilientstorage",
"name": "ResilientStorage"
}
]
},
"ansible_included_var_files": [
"/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/CentOS_8.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
"ansible_facts": {
"__ha_cluster_cloud_agents_packages": [
"resource-agents-aliyun",
"resource-agents-gcp",
"fence-agents-aliyun",
"fence-agents-aws",
"fence-agents-azure-arm",
"fence-agents-gce"
],
"__ha_cluster_repos": [
{
"id": "ha",
"name": "HighAvailability"
},
{
"id": "resilientstorage",
"name": "ResilientStorage"
}
]
},
"ansible_included_var_files": [
"/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/CentOS_8.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
"ansible_facts": {
"__ha_cluster_cloud_agents_packages": [
"resource-agents-aliyun",
"resource-agents-gcp",
"fence-agents-aliyun",
"fence-agents-aws",
"fence-agents-azure-arm",
"fence-agents-gce"
],
"__ha_cluster_repos": [
{
"id": "ha",
"name": "HighAvailability"
},
{
"id": "resilientstorage",
"name": "ResilientStorage"
}
]
},
"ansible_included_var_files": [
"/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/CentOS_8.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml"
}
TASK [fedora.linux_system_roles.ha_cluster : Set Linux Pacemaker shell specific variables] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:42
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.059) 0:00:02.014 ****
ok: [managed-node1] => {
"ansible_facts": {},
"ansible_included_var_files": [
"/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/shell_pcs.yml"
],
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Enable package repositories] ******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:6
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.017) 0:00:02.031 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Find platform/version specific tasks to enable repositories] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml:3
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.023) 0:00:02.055 ****
ok: [managed-node1] => (item=RedHat.yml) => {
"ansible_facts": {
"__ha_cluster_enable_repo_tasks_file": "/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/RedHat.yml"
},
"ansible_loop_var": "item",
"changed": false,
"item": "RedHat.yml"
}
ok: [managed-node1] => (item=CentOS.yml) => {
"ansible_facts": {
"__ha_cluster_enable_repo_tasks_file": "/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml"
},
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS.yml"
}
skipping: [managed-node1] => (item=CentOS_8.yml) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__ha_cluster_enable_repo_tasks_file_candidate is file",
"item": "CentOS_8.yml",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS_8.yml) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__ha_cluster_enable_repo_tasks_file_candidate is file",
"item": "CentOS_8.yml",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Run platform/version specific tasks to enable repositories] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml:21
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.046) 0:00:02.102 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : List active CentOS repositories] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml:3
Wednesday 11 December 2024 11:27:19 -0500 (0:00:00.072) 0:00:02.174 ****
ok: [managed-node1] => {
"changed": false,
"cmd": [
"dnf",
"repolist"
],
"delta": "0:00:00.219541",
"end": "2024-12-11 11:27:20.447074",
"rc": 0,
"start": "2024-12-11 11:27:20.227533"
}
STDOUT:
repo id repo name
appstream CentOS Stream 8 - AppStream
baseos CentOS Stream 8 - BaseOS
beaker-client Beaker Client - RedHatEnterpriseLinux8
beaker-harness Beaker harness
beakerlib-libraries Copr repo for beakerlib-libraries owned by bgoncalv
copr:copr.devel.redhat.com:lpol:qa-tools Copr repo for qa-tools owned by lpol
epel Extra Packages for Enterprise Linux 8 - x86_64
epel-debuginfo Extra Packages for Enterprise Linux 8 - x86_64 - Debug
epel-source Extra Packages for Enterprise Linux 8 - x86_64 - Source
extras CentOS Stream 8 - Extras
extras-common CentOS Stream 8 - Extras common packages
ha CentOS Stream 8 - HighAvailability
TASK [fedora.linux_system_roles.ha_cluster : Enable CentOS repositories] *******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml:10
Wednesday 11 December 2024 11:27:20 -0500 (0:00:00.787) 0:00:02.962 ****
skipping: [managed-node1] => (item={'id': 'ha', 'name': 'HighAvailability'}) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "item.id not in __ha_cluster_repolist.stdout",
"item": {
"id": "ha",
"name": "HighAvailability"
},
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item={'id': 'resilientstorage', 'name': 'ResilientStorage'}) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "item.name != \"ResilientStorage\" or ha_cluster_enable_repos_resilient_storage",
"item": {
"id": "resilientstorage",
"name": "ResilientStorage"
},
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => {
"changed": false
}
MSG:
All items skipped
TASK [fedora.linux_system_roles.ha_cluster : Install role essential packages] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:11
Wednesday 11 December 2024 11:27:20 -0500 (0:00:00.023) 0:00:02.985 ****
ok: [managed-node1] => {
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
TASK [fedora.linux_system_roles.ha_cluster : Check and prepare role variables] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:17
Wednesday 11 December 2024 11:27:23 -0500 (0:00:02.981) 0:00:05.967 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Discover cluster node names] ******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:3
Wednesday 11 December 2024 11:27:23 -0500 (0:00:00.059) 0:00:06.026 ****
ok: [managed-node1] => {
"ansible_facts": {
"__ha_cluster_node_name": "localhost"
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Collect cluster node names] *******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:7
Wednesday 11 December 2024 11:27:23 -0500 (0:00:00.055) 0:00:06.082 ****
ok: [managed-node1] => {
"ansible_facts": {
"__ha_cluster_all_node_names": [
"localhost"
]
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if ha_cluster_node_options contains unknown or duplicate nodes] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:16
Wednesday 11 December 2024 11:27:23 -0500 (0:00:00.059) 0:00:06.141 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "(\n __nodes_from_options != (__nodes_from_options | unique)\n) or (\n __nodes_from_options | difference(__ha_cluster_all_node_names)\n)\n",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Extract node options] *************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:30
Wednesday 11 December 2024 11:27:23 -0500 (0:00:00.038) 0:00:06.180 ****
ok: [managed-node1] => {
"ansible_facts": {
"__ha_cluster_local_node": {}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if passwords are not specified] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:43
Wednesday 11 December 2024 11:27:23 -0500 (0:00:00.045) 0:00:06.225 ****
skipping: [managed-node1] => (item=ha_cluster_hacluster_password) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "lookup(\"vars\", item, default=\"\") | string | length < 1",
"item": "ha_cluster_hacluster_password",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => {
"changed": false
}
MSG:
All items skipped
TASK [fedora.linux_system_roles.ha_cluster : Fail if nodes do not have the same number of SBD devices specified] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:53
Wednesday 11 December 2024 11:27:23 -0500 (0:00:00.068) 0:00:06.294 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_sbd_enabled",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if configuring qnetd on a cluster node] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:69
Wednesday 11 December 2024 11:27:23 -0500 (0:00:00.014) 0:00:06.309 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_qnetd.present | d(false)",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if no valid level is specified for a fencing level] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:79
Wednesday 11 December 2024 11:27:23 -0500 (0:00:00.036) 0:00:06.346 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if no target is specified for a fencing level] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:87
Wednesday 11 December 2024 11:27:23 -0500 (0:00:00.013) 0:00:06.359 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Extract qdevice settings] *********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:101
Wednesday 11 December 2024 11:27:23 -0500 (0:00:00.012) 0:00:06.371 ****
ok: [managed-node1] => {
"ansible_facts": {
"__ha_cluster_qdevice_host": "",
"__ha_cluster_qdevice_in_use": false,
"__ha_cluster_qdevice_model": "",
"__ha_cluster_qdevice_pcs_address": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Figure out if ATB needs to be enabled for SBD] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:110
Wednesday 11 December 2024 11:27:23 -0500 (0:00:00.046) 0:00:06.418 ****
ok: [managed-node1] => {
"ansible_facts": {
"__ha_cluster_sbd_needs_atb": false
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if SBD needs ATB enabled and the user configured ATB to be disabled] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:120
Wednesday 11 December 2024 11:27:24 -0500 (0:00:00.050) 0:00:06.469 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__ha_cluster_sbd_needs_atb | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if ha_cluster_pcsd_public_key_src and ha_cluster_pcsd_private_key_src are set along with ha_cluster_pcsd_certificates] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:127
Wednesday 11 December 2024 11:27:24 -0500 (0:00:00.090) 0:00:06.560 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_pcsd_public_key_src is not none",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch pcs capabilities] ***********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:141
Wednesday 11 December 2024 11:27:24 -0500 (0:00:00.063) 0:00:06.623 ****
ok: [managed-node1] => {
"changed": false,
"cmd": [
"pcs",
"--version",
"--full"
],
"delta": "0:00:00.514270",
"end": "2024-12-11 11:27:25.010649",
"rc": 0,
"start": "2024-12-11 11:27:24.496379"
}
STDOUT:
0.10.18
booth booth.enable-authfile.set booth.enable-authfile.unset cluster.config.backup-local cluster.config.restore-cluster cluster.config.restore-local cluster.config.uuid cluster.create cluster.create.enable cluster.create.local cluster.create.no-keys-sync cluster.create.separated-name-and-address cluster.create.start cluster.create.start.wait cluster.create.transport.knet cluster.create.transport.udp-udpu cluster.create.transport.udp-udpu.no-rrp cluster.destroy cluster.destroy.all cluster.report cluster.verify corosync.authkey.update corosync.config.get corosync.config.get.struct corosync.config.reload corosync.config.sync-to-local-cluster corosync.config.update corosync.link.add corosync.link.remove corosync.link.remove.list corosync.link.update corosync.qdevice corosync.qdevice.model.net corosync.quorum corosync.quorum.device corosync.quorum.device.heuristics corosync.quorum.device.model.net corosync.quorum.device.model.net.options_tls_and_kaptb corosync.quorum.set-expected-votes-runtime corosync.quorum.status corosync.quorum.unblock corosync.totem.block_unlisted_ips corosync.uidgid node.add node.add.enable node.add.separated-name-and-address node.add.start node.add.start.wait node.attributes node.attributes.set-list-for-node node.confirm-off node.fence node.guest node.kill node.maintenance node.maintenance.all node.maintenance.list node.maintenance.wait node.remote node.remote.onfail-demote node.remove node.remove-from-caches node.remove.list node.standby node.standby.all node.standby.list node.standby.wait node.start-stop-enable-disable node.start-stop-enable-disable.all node.start-stop-enable-disable.list node.start-stop-enable-disable.start-wait node.utilization node.utilization.set-list-for-node pcmk.acl.enable-disable pcmk.acl.group pcmk.acl.role pcmk.acl.role.create-with-permissions pcmk.acl.role.delete-with-users-groups pcmk.acl.user pcmk.alert pcmk.cib.checkpoints pcmk.cib.checkpoints.diff pcmk.cib.edit pcmk.cib.get pcmk.cib.get.scope pcmk.cib.roles.promoted-unpromoted pcmk.cib.set pcmk.constraint.colocation.set pcmk.constraint.colocation.set.options pcmk.constraint.colocation.simple pcmk.constraint.colocation.simple.options pcmk.constraint.hide-expired pcmk.constraint.location.simple pcmk.constraint.location.simple.options pcmk.constraint.location.simple.resource-regexp pcmk.constraint.location.simple.rule pcmk.constraint.location.simple.rule.node-attr-type-number pcmk.constraint.location.simple.rule.options pcmk.constraint.location.simple.rule.rule-add-remove pcmk.constraint.no-autocorrect pcmk.constraint.order.set pcmk.constraint.order.set.options pcmk.constraint.order.simple pcmk.constraint.order.simple.options pcmk.constraint.ticket.set pcmk.constraint.ticket.set.options pcmk.constraint.ticket.simple pcmk.constraint.ticket.simple.constraint-id pcmk.properties.cluster pcmk.properties.cluster.config.output-formats pcmk.properties.cluster.defaults pcmk.properties.cluster.describe pcmk.properties.cluster.describe.output-formats pcmk.properties.operation-defaults pcmk.properties.operation-defaults.multiple pcmk.properties.operation-defaults.rule pcmk.properties.operation-defaults.rule-rsc-op pcmk.properties.operation-defaults.rule.hide-expired pcmk.properties.operation-defaults.rule.node-attr-type-number pcmk.properties.resource-defaults pcmk.properties.resource-defaults.multiple pcmk.properties.resource-defaults.rule pcmk.properties.resource-defaults.rule-rsc-op pcmk.properties.resource-defaults.rule.hide-expired pcmk.properties.resource-defaults.rule.node-attr-type-number pcmk.resource.ban-move-clear pcmk.resource.ban-move-clear.bundles pcmk.resource.ban-move-clear.clear-expired pcmk.resource.ban-move-clear.clone pcmk.resource.bundle pcmk.resource.bundle.container-docker pcmk.resource.bundle.container-docker.promoted-max pcmk.resource.bundle.container-podman pcmk.resource.bundle.container-podman.promoted-max pcmk.resource.bundle.container-rkt pcmk.resource.bundle.container-rkt.promoted-max pcmk.resource.bundle.reset pcmk.resource.bundle.wait pcmk.resource.cleanup pcmk.resource.cleanup.one-resource pcmk.resource.cleanup.strict pcmk.resource.clone pcmk.resource.clone.custom-id pcmk.resource.clone.meta-in-create pcmk.resource.clone.wait pcmk.resource.config.output-formats pcmk.resource.create pcmk.resource.create.clone.custom-id pcmk.resource.create.in-existing-bundle pcmk.resource.create.meta pcmk.resource.create.no-master pcmk.resource.create.operations pcmk.resource.create.operations.onfail-demote pcmk.resource.create.promotable pcmk.resource.create.promotable.custom-id pcmk.resource.create.wait pcmk.resource.debug pcmk.resource.delete pcmk.resource.disable.safe pcmk.resource.disable.safe.brief pcmk.resource.disable.safe.tag pcmk.resource.disable.simulate pcmk.resource.disable.simulate.brief pcmk.resource.disable.simulate.tag pcmk.resource.enable-disable pcmk.resource.enable-disable.list pcmk.resource.enable-disable.tag pcmk.resource.enable-disable.wait pcmk.resource.failcount pcmk.resource.group pcmk.resource.group.add-remove-list pcmk.resource.group.wait pcmk.resource.manage-unmanage pcmk.resource.manage-unmanage.list pcmk.resource.manage-unmanage.tag pcmk.resource.manage-unmanage.with-monitor pcmk.resource.move.autoclean pcmk.resource.promotable pcmk.resource.promotable.custom-id pcmk.resource.promotable.meta-in-create pcmk.resource.promotable.wait pcmk.resource.refresh pcmk.resource.refresh.one-resource pcmk.resource.refresh.strict pcmk.resource.relations pcmk.resource.relocate pcmk.resource.restart pcmk.resource.update pcmk.resource.update-meta pcmk.resource.update-meta.list pcmk.resource.update-meta.wait pcmk.resource.update-operations pcmk.resource.update-operations.onfail-demote pcmk.resource.update.meta pcmk.resource.update.operations pcmk.resource.update.operations.onfail-demote pcmk.resource.update.wait pcmk.resource.utilization pcmk.resource.utilization-set-list-for-resource pcmk.stonith.cleanup pcmk.stonith.cleanup.one-resource pcmk.stonith.cleanup.strict pcmk.stonith.create pcmk.stonith.create.in-group pcmk.stonith.create.meta pcmk.stonith.create.operations pcmk.stonith.create.operations.onfail-demote pcmk.stonith.create.wait pcmk.stonith.delete pcmk.stonith.enable-disable pcmk.stonith.enable-disable.list pcmk.stonith.enable-disable.wait pcmk.stonith.history.cleanup pcmk.stonith.history.show pcmk.stonith.history.update pcmk.stonith.levels pcmk.stonith.levels.add-remove-devices-list pcmk.stonith.levels.clear pcmk.stonith.levels.node-attr pcmk.stonith.levels.node-regexp pcmk.stonith.levels.verify pcmk.stonith.refresh pcmk.stonith.refresh.one-resource pcmk.stonith.refresh.strict pcmk.stonith.update pcmk.stonith.update.scsi-devices pcmk.stonith.update.scsi-devices.add-remove pcmk.stonith.update.scsi-devices.mpath pcmk.tag pcmk.tag.resources pcs.auth.client pcs.auth.client.cluster pcs.auth.client.token pcs.auth.deauth-client pcs.auth.deauth-server pcs.auth.no-bidirectional pcs.auth.separated-name-and-address pcs.auth.server.token pcs.cfg-in-file.cib pcs.daemon-ssl-cert.set pcs.daemon-ssl-cert.sync-to-local-cluster pcs.disaster-recovery.essentials pcs.request-timeout resource-agents.describe resource-agents.list resource-agents.list.detailed resource-agents.ocf.version-1-0 resource-agents.ocf.version-1-1 resource-agents.self-validation sbd sbd.option-timeout-action sbd.shared-block-device status.corosync.membership status.pcmk.resources.hide-inactive status.pcmk.resources.id status.pcmk.resources.node status.pcmk.resources.orphaned status.pcmk.xml stonith-agents.describe stonith-agents.list stonith-agents.list.detailed stonith-agents.ocf.version-1-0 stonith-agents.ocf.version-1-1 stonith-agents.self-validation
TASK [fedora.linux_system_roles.ha_cluster : Parse pcs capabilities] ***********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:148
Wednesday 11 December 2024 11:27:25 -0500 (0:00:00.902) 0:00:07.525 ****
ok: [managed-node1] => {
"ansible_facts": {
"__ha_cluster_pcs_capabilities": [
"booth",
"booth.enable-authfile.set",
"booth.enable-authfile.unset",
"cluster.config.backup-local",
"cluster.config.restore-cluster",
"cluster.config.restore-local",
"cluster.config.uuid",
"cluster.create",
"cluster.create.enable",
"cluster.create.local",
"cluster.create.no-keys-sync",
"cluster.create.separated-name-and-address",
"cluster.create.start",
"cluster.create.start.wait",
"cluster.create.transport.knet",
"cluster.create.transport.udp-udpu",
"cluster.create.transport.udp-udpu.no-rrp",
"cluster.destroy",
"cluster.destroy.all",
"cluster.report",
"cluster.verify",
"corosync.authkey.update",
"corosync.config.get",
"corosync.config.get.struct",
"corosync.config.reload",
"corosync.config.sync-to-local-cluster",
"corosync.config.update",
"corosync.link.add",
"corosync.link.remove",
"corosync.link.remove.list",
"corosync.link.update",
"corosync.qdevice",
"corosync.qdevice.model.net",
"corosync.quorum",
"corosync.quorum.device",
"corosync.quorum.device.heuristics",
"corosync.quorum.device.model.net",
"corosync.quorum.device.model.net.options_tls_and_kaptb",
"corosync.quorum.set-expected-votes-runtime",
"corosync.quorum.status",
"corosync.quorum.unblock",
"corosync.totem.block_unlisted_ips",
"corosync.uidgid",
"node.add",
"node.add.enable",
"node.add.separated-name-and-address",
"node.add.start",
"node.add.start.wait",
"node.attributes",
"node.attributes.set-list-for-node",
"node.confirm-off",
"node.fence",
"node.guest",
"node.kill",
"node.maintenance",
"node.maintenance.all",
"node.maintenance.list",
"node.maintenance.wait",
"node.remote",
"node.remote.onfail-demote",
"node.remove",
"node.remove-from-caches",
"node.remove.list",
"node.standby",
"node.standby.all",
"node.standby.list",
"node.standby.wait",
"node.start-stop-enable-disable",
"node.start-stop-enable-disable.all",
"node.start-stop-enable-disable.list",
"node.start-stop-enable-disable.start-wait",
"node.utilization",
"node.utilization.set-list-for-node",
"pcmk.acl.enable-disable",
"pcmk.acl.group",
"pcmk.acl.role",
"pcmk.acl.role.create-with-permissions",
"pcmk.acl.role.delete-with-users-groups",
"pcmk.acl.user",
"pcmk.alert",
"pcmk.cib.checkpoints",
"pcmk.cib.checkpoints.diff",
"pcmk.cib.edit",
"pcmk.cib.get",
"pcmk.cib.get.scope",
"pcmk.cib.roles.promoted-unpromoted",
"pcmk.cib.set",
"pcmk.constraint.colocation.set",
"pcmk.constraint.colocation.set.options",
"pcmk.constraint.colocation.simple",
"pcmk.constraint.colocation.simple.options",
"pcmk.constraint.hide-expired",
"pcmk.constraint.location.simple",
"pcmk.constraint.location.simple.options",
"pcmk.constraint.location.simple.resource-regexp",
"pcmk.constraint.location.simple.rule",
"pcmk.constraint.location.simple.rule.node-attr-type-number",
"pcmk.constraint.location.simple.rule.options",
"pcmk.constraint.location.simple.rule.rule-add-remove",
"pcmk.constraint.no-autocorrect",
"pcmk.constraint.order.set",
"pcmk.constraint.order.set.options",
"pcmk.constraint.order.simple",
"pcmk.constraint.order.simple.options",
"pcmk.constraint.ticket.set",
"pcmk.constraint.ticket.set.options",
"pcmk.constraint.ticket.simple",
"pcmk.constraint.ticket.simple.constraint-id",
"pcmk.properties.cluster",
"pcmk.properties.cluster.config.output-formats",
"pcmk.properties.cluster.defaults",
"pcmk.properties.cluster.describe",
"pcmk.properties.cluster.describe.output-formats",
"pcmk.properties.operation-defaults",
"pcmk.properties.operation-defaults.multiple",
"pcmk.properties.operation-defaults.rule",
"pcmk.properties.operation-defaults.rule-rsc-op",
"pcmk.properties.operation-defaults.rule.hide-expired",
"pcmk.properties.operation-defaults.rule.node-attr-type-number",
"pcmk.properties.resource-defaults",
"pcmk.properties.resource-defaults.multiple",
"pcmk.properties.resource-defaults.rule",
"pcmk.properties.resource-defaults.rule-rsc-op",
"pcmk.properties.resource-defaults.rule.hide-expired",
"pcmk.properties.resource-defaults.rule.node-attr-type-number",
"pcmk.resource.ban-move-clear",
"pcmk.resource.ban-move-clear.bundles",
"pcmk.resource.ban-move-clear.clear-expired",
"pcmk.resource.ban-move-clear.clone",
"pcmk.resource.bundle",
"pcmk.resource.bundle.container-docker",
"pcmk.resource.bundle.container-docker.promoted-max",
"pcmk.resource.bundle.container-podman",
"pcmk.resource.bundle.container-podman.promoted-max",
"pcmk.resource.bundle.container-rkt",
"pcmk.resource.bundle.container-rkt.promoted-max",
"pcmk.resource.bundle.reset",
"pcmk.resource.bundle.wait",
"pcmk.resource.cleanup",
"pcmk.resource.cleanup.one-resource",
"pcmk.resource.cleanup.strict",
"pcmk.resource.clone",
"pcmk.resource.clone.custom-id",
"pcmk.resource.clone.meta-in-create",
"pcmk.resource.clone.wait",
"pcmk.resource.config.output-formats",
"pcmk.resource.create",
"pcmk.resource.create.clone.custom-id",
"pcmk.resource.create.in-existing-bundle",
"pcmk.resource.create.meta",
"pcmk.resource.create.no-master",
"pcmk.resource.create.operations",
"pcmk.resource.create.operations.onfail-demote",
"pcmk.resource.create.promotable",
"pcmk.resource.create.promotable.custom-id",
"pcmk.resource.create.wait",
"pcmk.resource.debug",
"pcmk.resource.delete",
"pcmk.resource.disable.safe",
"pcmk.resource.disable.safe.brief",
"pcmk.resource.disable.safe.tag",
"pcmk.resource.disable.simulate",
"pcmk.resource.disable.simulate.brief",
"pcmk.resource.disable.simulate.tag",
"pcmk.resource.enable-disable",
"pcmk.resource.enable-disable.list",
"pcmk.resource.enable-disable.tag",
"pcmk.resource.enable-disable.wait",
"pcmk.resource.failcount",
"pcmk.resource.group",
"pcmk.resource.group.add-remove-list",
"pcmk.resource.group.wait",
"pcmk.resource.manage-unmanage",
"pcmk.resource.manage-unmanage.list",
"pcmk.resource.manage-unmanage.tag",
"pcmk.resource.manage-unmanage.with-monitor",
"pcmk.resource.move.autoclean",
"pcmk.resource.promotable",
"pcmk.resource.promotable.custom-id",
"pcmk.resource.promotable.meta-in-create",
"pcmk.resource.promotable.wait",
"pcmk.resource.refresh",
"pcmk.resource.refresh.one-resource",
"pcmk.resource.refresh.strict",
"pcmk.resource.relations",
"pcmk.resource.relocate",
"pcmk.resource.restart",
"pcmk.resource.update",
"pcmk.resource.update-meta",
"pcmk.resource.update-meta.list",
"pcmk.resource.update-meta.wait",
"pcmk.resource.update-operations",
"pcmk.resource.update-operations.onfail-demote",
"pcmk.resource.update.meta",
"pcmk.resource.update.operations",
"pcmk.resource.update.operations.onfail-demote",
"pcmk.resource.update.wait",
"pcmk.resource.utilization",
"pcmk.resource.utilization-set-list-for-resource",
"pcmk.stonith.cleanup",
"pcmk.stonith.cleanup.one-resource",
"pcmk.stonith.cleanup.strict",
"pcmk.stonith.create",
"pcmk.stonith.create.in-group",
"pcmk.stonith.create.meta",
"pcmk.stonith.create.operations",
"pcmk.stonith.create.operations.onfail-demote",
"pcmk.stonith.create.wait",
"pcmk.stonith.delete",
"pcmk.stonith.enable-disable",
"pcmk.stonith.enable-disable.list",
"pcmk.stonith.enable-disable.wait",
"pcmk.stonith.history.cleanup",
"pcmk.stonith.history.show",
"pcmk.stonith.history.update",
"pcmk.stonith.levels",
"pcmk.stonith.levels.add-remove-devices-list",
"pcmk.stonith.levels.clear",
"pcmk.stonith.levels.node-attr",
"pcmk.stonith.levels.node-regexp",
"pcmk.stonith.levels.verify",
"pcmk.stonith.refresh",
"pcmk.stonith.refresh.one-resource",
"pcmk.stonith.refresh.strict",
"pcmk.stonith.update",
"pcmk.stonith.update.scsi-devices",
"pcmk.stonith.update.scsi-devices.add-remove",
"pcmk.stonith.update.scsi-devices.mpath",
"pcmk.tag",
"pcmk.tag.resources",
"pcs.auth.client",
"pcs.auth.client.cluster",
"pcs.auth.client.token",
"pcs.auth.deauth-client",
"pcs.auth.deauth-server",
"pcs.auth.no-bidirectional",
"pcs.auth.separated-name-and-address",
"pcs.auth.server.token",
"pcs.cfg-in-file.cib",
"pcs.daemon-ssl-cert.set",
"pcs.daemon-ssl-cert.sync-to-local-cluster",
"pcs.disaster-recovery.essentials",
"pcs.request-timeout",
"resource-agents.describe",
"resource-agents.list",
"resource-agents.list.detailed",
"resource-agents.ocf.version-1-0",
"resource-agents.ocf.version-1-1",
"resource-agents.self-validation",
"sbd",
"sbd.option-timeout-action",
"sbd.shared-block-device",
"status.corosync.membership",
"status.pcmk.resources.hide-inactive",
"status.pcmk.resources.id",
"status.pcmk.resources.node",
"status.pcmk.resources.orphaned",
"status.pcmk.xml",
"stonith-agents.describe",
"stonith-agents.list",
"stonith-agents.list.detailed",
"stonith-agents.ocf.version-1-0",
"stonith-agents.ocf.version-1-1",
"stonith-agents.self-validation"
],
"__ha_cluster_pcsd_capabilities_available": false
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch pcsd capabilities] **********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:155
Wednesday 11 December 2024 11:27:25 -0500 (0:00:00.029) 0:00:07.555 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__ha_cluster_pcsd_capabilities_available",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Parse pcsd capabilities] **********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:163
Wednesday 11 December 2024 11:27:25 -0500 (0:00:00.015) 0:00:07.571 ****
ok: [managed-node1] => {
"ansible_facts": {
"__ha_cluster_pcsd_capabilities": []
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if pcs is to old to configure resources and operations defaults] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:172
Wednesday 11 December 2024 11:27:25 -0500 (0:00:00.039) 0:00:07.611 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "( ha_cluster_resource_defaults and not 'pcmk.properties.resource-defaults.multiple' in __ha_cluster_pcs_capabilities ) or ( ha_cluster_resource_operation_defaults and not 'pcmk.properties.operation-defaults.multiple' in __ha_cluster_pcs_capabilities )",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Generate a password hash] *********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:26
Wednesday 11 December 2024 11:27:25 -0500 (0:00:00.018) 0:00:07.629 ****
ok: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Set hacluster password] ***********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:38
Wednesday 11 December 2024 11:27:25 -0500 (0:00:00.438) 0:00:08.067 ****
ok: [managed-node1] => {
"append": false,
"changed": false,
"comment": "cluster user",
"group": 189,
"home": "/home/hacluster",
"move_home": false,
"name": "hacluster",
"password": "NOT_LOGGING_PASSWORD",
"shell": "/sbin/nologin",
"state": "present",
"uid": 189
}
TASK [fedora.linux_system_roles.ha_cluster : Configure shell] ******************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:43
Wednesday 11 December 2024 11:27:26 -0500 (0:00:00.555) 0:00:08.623 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Stop pcsd] ************************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:6
Wednesday 11 December 2024 11:27:26 -0500 (0:00:00.038) 0:00:08.662 ****
changed: [managed-node1] => {
"changed": true,
"name": "pcsd",
"state": "stopped",
"status": {
"ActiveEnterTimestamp": "Wed 2024-12-11 11:25:25 EST",
"ActiveEnterTimestampMonotonic": "951639778",
"ActiveExitTimestamp": "Wed 2024-12-11 11:25:23 EST",
"ActiveExitTimestampMonotonic": "949421954",
"ActiveState": "active",
"After": "basic.target sysinit.target system.slice network-online.target pcsd-ruby.service systemd-journald.socket",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-12-11 11:25:25 EST",
"AssertTimestampMonotonic": "951369073",
"Before": "shutdown.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-12-11 11:25:25 EST",
"ConditionTimestampMonotonic": "951369071",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ConsistsOf": "pcsd-ruby.service",
"ControlGroup": "/system.slice/pcsd.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "PCS GUI and remote configuration interface",
"DevicePolicy": "auto",
"Documentation": "man:pcsd(8) man:pcs(8)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/pcsd (ignore_errors=no)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "66245",
"ExecMainStartTimestamp": "Wed 2024-12-11 11:25:25 EST",
"ExecMainStartTimestampMonotonic": "951370208",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pcsd.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pcsd.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Wed 2024-12-11 11:25:23 EST",
"InactiveEnterTimestampMonotonic": "949454034",
"InactiveExitTimestamp": "Wed 2024-12-11 11:25:25 EST",
"InactiveExitTimestampMonotonic": "951370289",
"InvocationID": "2a5e3f8a1d15487e88829f210cc0bf07",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "66245",
"MemoryAccounting": "yes",
"MemoryCurrent": "23797760",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pcsd.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target system.slice network-online.target pcsd-ruby.service",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Wed 2024-12-11 11:25:25 EST",
"StateChangeTimestampMonotonic": "951639778",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "1",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestamp": "Wed 2024-12-11 11:25:25 EST",
"WatchdogTimestampMonotonic": "951639775",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Regenerate pcsd TLS certificate and key] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:11
Wednesday 11 December 2024 11:27:26 -0500 (0:00:00.774) 0:00:09.436 ****
skipping: [managed-node1] => (item=/var/lib/pcsd/pcsd.key) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "ha_cluster_regenerate_keys",
"item": "/var/lib/pcsd/pcsd.key",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=/var/lib/pcsd/pcsd.crt) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "ha_cluster_regenerate_keys",
"item": "/var/lib/pcsd/pcsd.crt",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => {
"changed": false
}
MSG:
All items skipped
TASK [fedora.linux_system_roles.ha_cluster : Get the stat of /var/lib/pcsd] ****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:25
Wednesday 11 December 2024 11:27:26 -0500 (0:00:00.019) 0:00:09.456 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_pcsd_certificates | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Allow certmonger to write into pcsd's certificate directory] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:30
Wednesday 11 December 2024 11:27:27 -0500 (0:00:00.037) 0:00:09.493 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_pcsd_certificates | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [Ensure the name of ha_cluster_pcsd_certificates is /var/lib/pcsd/pcsd; Create certificates using the certificate role] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:37
Wednesday 11 December 2024 11:27:27 -0500 (0:00:00.037) 0:00:09.530 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_pcsd_certificates | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Set pcsd's certificate directory back to cluster_var_lib_t] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:49
Wednesday 11 December 2024 11:27:27 -0500 (0:00:00.039) 0:00:09.570 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_pcsd_certificates | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute pcsd TLS private key] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:64
Wednesday 11 December 2024 11:27:27 -0500 (0:00:00.036) 0:00:09.607 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_pcsd_public_key_src is string",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute pcsd TLS certificate] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:71
Wednesday 11 December 2024 11:27:27 -0500 (0:00:00.048) 0:00:09.656 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_pcsd_public_key_src is string",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute pcs_settings.conf] *****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:79
Wednesday 11 December 2024 11:27:27 -0500 (0:00:00.055) 0:00:09.712 ****
ok: [managed-node1] => {
"changed": false,
"checksum": "9bbea634a798cf0976b80cd3c4e34aca6a6a7d44",
"dest": "/var/lib/pcsd/pcs_settings.conf",
"gid": 0,
"group": "root",
"mode": "0644",
"owner": "root",
"path": "/var/lib/pcsd/pcs_settings.conf",
"secontext": "system_u:object_r:cluster_var_lib_t:s0",
"size": 361,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Start pcsd with updated config files and configure it to start on boot] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:88
Wednesday 11 December 2024 11:27:28 -0500 (0:00:00.772) 0:00:10.485 ****
changed: [managed-node1] => {
"changed": true,
"enabled": true,
"name": "pcsd",
"state": "started",
"status": {
"ActiveEnterTimestamp": "Wed 2024-12-11 11:25:25 EST",
"ActiveEnterTimestampMonotonic": "951639778",
"ActiveExitTimestamp": "Wed 2024-12-11 11:27:26 EST",
"ActiveExitTimestampMonotonic": "1072875982",
"ActiveState": "inactive",
"After": "basic.target sysinit.target system.slice network-online.target pcsd-ruby.service systemd-journald.socket",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-12-11 11:25:25 EST",
"AssertTimestampMonotonic": "951369073",
"Before": "shutdown.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-12-11 11:25:25 EST",
"ConditionTimestampMonotonic": "951369071",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ConsistsOf": "pcsd-ruby.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "PCS GUI and remote configuration interface",
"DevicePolicy": "auto",
"Documentation": "man:pcsd(8) man:pcs(8)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/pcsd (ignore_errors=no)",
"ExecMainCode": "1",
"ExecMainExitTimestamp": "Wed 2024-12-11 11:27:26 EST",
"ExecMainExitTimestampMonotonic": "1072907439",
"ExecMainPID": "66245",
"ExecMainStartTimestamp": "Wed 2024-12-11 11:25:25 EST",
"ExecMainStartTimestampMonotonic": "951370208",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; ignore_errors=no ; start_time=[Wed 2024-12-11 11:25:25 EST] ; stop_time=[Wed 2024-12-11 11:27:26 EST] ; pid=66245 ; code=exited ; status=0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pcsd.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pcsd.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Wed 2024-12-11 11:27:26 EST",
"InactiveEnterTimestampMonotonic": "1072907541",
"InactiveExitTimestamp": "Wed 2024-12-11 11:25:25 EST",
"InactiveExitTimestampMonotonic": "951370289",
"InvocationID": "2a5e3f8a1d15487e88829f210cc0bf07",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pcsd.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target system.slice network-online.target pcsd-ruby.service",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Wed 2024-12-11 11:27:26 EST",
"StateChangeTimestampMonotonic": "1072907541",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Configure firewall] ***************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:49
Wednesday 11 December 2024 11:27:29 -0500 (0:00:01.180) 0:00:11.665 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/firewall.yml for managed-node1
TASK [Ensure the service and the ports status with the firewall role] **********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/firewall.yml:3
Wednesday 11 December 2024 11:27:29 -0500 (0:00:00.042) 0:00:11.708 ****
TASK [fedora.linux_system_roles.firewall : Setup firewalld] ********************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2
Wednesday 11 December 2024 11:27:29 -0500 (0:00:00.092) 0:00:11.800 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed-node1
TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2
Wednesday 11 December 2024 11:27:29 -0500 (0:00:00.073) 0:00:11.874 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__firewall_required_facts | difference(ansible_facts.keys() | list) | length > 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Check if system is ostree] **********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10
Wednesday 11 December 2024 11:27:29 -0500 (0:00:00.044) 0:00:11.919 ****
ok: [managed-node1] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15
Wednesday 11 December 2024 11:27:29 -0500 (0:00:00.349) 0:00:12.269 ****
ok: [managed-node1] => {
"ansible_facts": {
"__firewall_is_ostree": false
},
"changed": false
}
TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22
Wednesday 11 December 2024 11:27:29 -0500 (0:00:00.044) 0:00:12.313 ****
ok: [managed-node1] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27
Wednesday 11 December 2024 11:27:30 -0500 (0:00:00.343) 0:00:12.657 ****
ok: [managed-node1] => {
"ansible_facts": {
"__firewall_is_transactional": false
},
"changed": false
}
TASK [fedora.linux_system_roles.firewall : Install firewalld] ******************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31
Wednesday 11 December 2024 11:27:30 -0500 (0:00:00.045) 0:00:12.702 ****
ok: [managed-node1] => {
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43
Wednesday 11 December 2024 11:27:33 -0500 (0:00:02.941) 0:00:15.644 ****
skipping: [managed-node1] => {
"false_condition": "__firewall_is_transactional | d(false)"
}
TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48
Wednesday 11 December 2024 11:27:33 -0500 (0:00:00.038) 0:00:15.682 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__firewall_is_transactional | d(false)",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53
Wednesday 11 December 2024 11:27:33 -0500 (0:00:00.038) 0:00:15.721 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__firewall_is_transactional | d(false)",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Collect service facts] **************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5
Wednesday 11 December 2024 11:27:33 -0500 (0:00:00.040) 0:00:15.762 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "firewall_disable_conflicting_services | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9
Wednesday 11 December 2024 11:27:33 -0500 (0:00:00.040) 0:00:15.802 ****
skipping: [managed-node1] => (item=nftables) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall_disable_conflicting_services | bool",
"item": "nftables",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=iptables) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall_disable_conflicting_services | bool",
"item": "iptables",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=ufw) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall_disable_conflicting_services | bool",
"item": "ufw",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => {
"changed": false
}
MSG:
All items skipped
TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] ***********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22
Wednesday 11 December 2024 11:27:33 -0500 (0:00:00.055) 0:00:15.858 ****
ok: [managed-node1] => {
"changed": false,
"name": "firewalld",
"status": {
"ActiveEnterTimestamp": "Wed 2024-12-11 11:25:31 EST",
"ActiveEnterTimestampMonotonic": "957168078",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "basic.target dbus.service sysinit.target dbus.socket polkit.service system.slice",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-12-11 11:25:30 EST",
"AssertTimestampMonotonic": "956852274",
"Before": "multi-user.target shutdown.target network-pre.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"BusName": "org.fedoraproject.FirewallD1",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "yes",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-12-11 11:25:30 EST",
"ConditionTimestampMonotonic": "956852273",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target ebtables.service ipset.service ip6tables.service iptables.service nftables.service",
"ControlGroup": "/system.slice/firewalld.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "firewalld - dynamic firewall daemon",
"DevicePolicy": "auto",
"Documentation": "man:firewalld(1)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "66904",
"ExecMainStartTimestamp": "Wed 2024-12-11 11:25:30 EST",
"ExecMainStartTimestampMonotonic": "956853896",
"ExecMainStatus": "0",
"ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/firewalld.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "firewalld.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Wed 2024-12-11 11:25:30 EST",
"InactiveExitTimestampMonotonic": "956853999",
"InvocationID": "bb0ae9b3399c47dfa231eb5467294a48",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "66904",
"MemoryAccounting": "yes",
"MemoryCurrent": "41226240",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "firewalld.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target dbus.socket system.slice",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "null",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Wed 2024-12-11 11:25:31 EST",
"StateChangeTimestampMonotonic": "957168078",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "2",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "dbus",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "enabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "network-pre.target",
"WatchdogTimestamp": "Wed 2024-12-11 11:25:31 EST",
"WatchdogTimestampMonotonic": "957168075",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28
Wednesday 11 December 2024 11:27:33 -0500 (0:00:00.516) 0:00:16.375 ****
ok: [managed-node1] => {
"changed": false,
"enabled": true,
"name": "firewalld",
"state": "started",
"status": {
"ActiveEnterTimestamp": "Wed 2024-12-11 11:25:31 EST",
"ActiveEnterTimestampMonotonic": "957168078",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "basic.target dbus.service sysinit.target dbus.socket polkit.service system.slice",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-12-11 11:25:30 EST",
"AssertTimestampMonotonic": "956852274",
"Before": "multi-user.target shutdown.target network-pre.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"BusName": "org.fedoraproject.FirewallD1",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "yes",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-12-11 11:25:30 EST",
"ConditionTimestampMonotonic": "956852273",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target ebtables.service ipset.service ip6tables.service iptables.service nftables.service",
"ControlGroup": "/system.slice/firewalld.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "firewalld - dynamic firewall daemon",
"DevicePolicy": "auto",
"Documentation": "man:firewalld(1)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/firewalld (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "66904",
"ExecMainStartTimestamp": "Wed 2024-12-11 11:25:30 EST",
"ExecMainStartTimestampMonotonic": "956853896",
"ExecMainStatus": "0",
"ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/firewalld.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "firewalld.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Wed 2024-12-11 11:25:30 EST",
"InactiveExitTimestampMonotonic": "956853999",
"InvocationID": "bb0ae9b3399c47dfa231eb5467294a48",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "mixed",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "66904",
"MemoryAccounting": "yes",
"MemoryCurrent": "41226240",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "firewalld.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target dbus.socket system.slice",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "null",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Wed 2024-12-11 11:25:31 EST",
"StateChangeTimestampMonotonic": "957168078",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "2",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "dbus",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "enabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "network-pre.target",
"WatchdogTimestamp": "Wed 2024-12-11 11:25:31 EST",
"WatchdogTimestampMonotonic": "957168075",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34
Wednesday 11 December 2024 11:27:34 -0500 (0:00:00.495) 0:00:16.870 ****
ok: [managed-node1] => {
"ansible_facts": {
"__firewall_previous_replaced": false,
"__firewall_python_cmd": "/usr/libexec/platform-python",
"__firewall_report_changed": true
},
"changed": false
}
TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43
Wednesday 11 December 2024 11:27:34 -0500 (0:00:00.099) 0:00:16.970 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__firewall_previous_replaced | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55
Wednesday 11 December 2024 11:27:34 -0500 (0:00:00.046) 0:00:17.016 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__firewall_previous_replaced | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Configure firewall] *****************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71
Wednesday 11 December 2024 11:27:34 -0500 (0:00:00.060) 0:00:17.077 ****
ok: [managed-node1] => (item={'service': 'high-availability', 'state': 'enabled'}) => {
"__firewall_changed": false,
"ansible_loop_var": "item",
"changed": false,
"item": {
"service": "high-availability",
"state": "enabled"
}
}
ok: [managed-node1] => (item={'port': '1229/tcp', 'state': 'enabled'}) => {
"__firewall_changed": false,
"ansible_loop_var": "item",
"changed": false,
"item": {
"port": "1229/tcp",
"state": "enabled"
}
}
TASK [fedora.linux_system_roles.firewall : Gather firewall config information] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120
Wednesday 11 December 2024 11:27:35 -0500 (0:00:01.192) 0:00:18.269 ****
skipping: [managed-node1] => (item={'service': 'high-availability', 'state': 'enabled'}) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall | length == 1",
"item": {
"service": "high-availability",
"state": "enabled"
},
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item={'port': '1229/tcp', 'state': 'enabled'}) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "firewall | length == 1",
"item": {
"port": "1229/tcp",
"state": "enabled"
},
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => {
"changed": false
}
MSG:
All items skipped
TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] *******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130
Wednesday 11 December 2024 11:27:35 -0500 (0:00:00.084) 0:00:18.353 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "firewall | length == 1",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139
Wednesday 11 December 2024 11:27:35 -0500 (0:00:00.051) 0:00:18.404 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "firewall == None or firewall | length == 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] *******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144
Wednesday 11 December 2024 11:27:35 -0500 (0:00:00.045) 0:00:18.449 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "firewall == None or firewall | length == 0",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153
Wednesday 11 December 2024 11:27:36 -0500 (0:00:00.044) 0:00:18.494 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__firewall_previous_replaced | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Calculate what has changed] *********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163
Wednesday 11 December 2024 11:27:36 -0500 (0:00:00.037) 0:00:18.532 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__firewall_previous_replaced | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.firewall : Show diffs] *************************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169
Wednesday 11 December 2024 11:27:36 -0500 (0:00:00.039) 0:00:18.572 ****
skipping: [managed-node1] => {
"false_condition": "__firewall_previous_replaced | bool"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure selinux] ****************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:52
Wednesday 11 December 2024 11:27:36 -0500 (0:00:00.053) 0:00:18.625 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Populate service facts] ***********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:3
Wednesday 11 December 2024 11:27:36 -0500 (0:00:00.054) 0:00:18.679 ****
ok: [managed-node1] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"auto-cpufreq.service": {
"name": "auto-cpufreq.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qdevice.service": {
"name": "corosync-qdevice.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"display-manager.service": {
"name": "display-manager.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"fcoe.service": {
"name": "fcoe.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"hv_kvp_daemon.service": {
"name": "hv_kvp_daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ip6tables.service": {
"name": "ip6tables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ipset.service": {
"name": "ipset.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iptables.service": {
"name": "iptables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"iscsi-shutdown.service": {
"name": "iscsi-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iscsi.service": {
"name": "iscsi.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iscsid.service": {
"name": "iscsid.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-activation-early.service": {
"name": "lvm2-activation-early.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"lvm2-activation.service": {
"name": "lvm2-activation.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"modprobe@efi_pstore.service": {
"name": "modprobe@efi_pstore.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"network.service": {
"name": "network.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"ntpd.service": {
"name": "ntpd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ntpdate.service": {
"name": "ntpdate.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"power-profiles-daemon.service": {
"name": "power-profiles-daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rbdmap.service": {
"name": "rbdmap.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"snapd.seeded.service": {
"name": "snapd.seeded.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sntp.service": {
"name": "sntp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen.service": {
"name": "sshd-keygen.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-networkd-wait-online.service": {
"name": "systemd-networkd-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-timesyncd.service": {
"name": "systemd-timesyncd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tlp.service": {
"name": "tlp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"ypbind.service": {
"name": "ypbind.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"yppasswdd.service": {
"name": "yppasswdd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ypserv.service": {
"name": "ypserv.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ypxfrd.service": {
"name": "ypxfrd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Set the fence-virt/fence-agents port to _ha_cluster_selinux] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:12
Wednesday 11 December 2024 11:27:37 -0500 (0:00:01.673) 0:00:20.353 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_manage_selinux | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Get associated selinux ports] *****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:20
Wednesday 11 December 2024 11:27:37 -0500 (0:00:00.044) 0:00:20.398 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_manage_selinux | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Add the high-availability service ports to _ha_cluster_selinux] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:28
Wednesday 11 December 2024 11:27:37 -0500 (0:00:00.044) 0:00:20.442 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_manage_selinux | bool",
"skip_reason": "Conditional result was False"
}
TASK [Ensure the service and the ports status with the selinux role] ***********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:35
Wednesday 11 December 2024 11:27:38 -0500 (0:00:00.044) 0:00:20.487 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_manage_selinux | bool",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Install cluster packages] *********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:58
Wednesday 11 December 2024 11:27:38 -0500 (0:00:00.045) 0:00:20.533 ****
ok: [managed-node1] => {
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
TASK [fedora.linux_system_roles.ha_cluster : Distribute fence-virt authkey] ****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:64
Wednesday 11 December 2024 11:27:40 -0500 (0:00:02.868) 0:00:23.401 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/distribute-fence-virt-key.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Create /etc/cluster directory] ****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/distribute-fence-virt-key.yml:3
Wednesday 11 December 2024 11:27:40 -0500 (0:00:00.051) 0:00:23.452 ****
ok: [managed-node1] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/cluster",
"secontext": "unconfined_u:object_r:etc_t:s0",
"size": 27,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Get fence_xvm.key] ****************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/distribute-fence-virt-key.yml:9
Wednesday 11 December 2024 11:27:41 -0500 (0:00:00.363) 0:00:23.815 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on the controller fence_xvm.key] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:16
Wednesday 11 December 2024 11:27:41 -0500 (0:00:00.058) 0:00:23.873 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "preshared_key_src is string and preshared_key_src | length > 1",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from the controller fence_xvm.key] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:21
Wednesday 11 December 2024 11:27:41 -0500 (0:00:00.049) 0:00:23.923 ****
skipping: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from the controller fence_xvm.key] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:29
Wednesday 11 December 2024 11:27:41 -0500 (0:00:00.056) 0:00:23.980 ****
skipping: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Generate key using OpenSSL fence_xvm.key] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:44
Wednesday 11 December 2024 11:27:41 -0500 (0:00:00.060) 0:00:24.040 ****
ok: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch generated fence_xvm.key] ****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:50
Wednesday 11 December 2024 11:27:41 -0500 (0:00:00.355) 0:00:24.396 ****
ok: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on cluster nodes fence_xvm.key] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:60
Wednesday 11 December 2024 11:27:41 -0500 (0:00:00.048) 0:00:24.445 ****
ok: [managed-node1] => {
"changed": false,
"stat": {
"atime": 1733933852.9324293,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "binary",
"checksum": "fc07c1393356f429049baf24c4f46cce9698df8e",
"ctime": 1733933712.1937292,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 337641626,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/octet-stream",
"mode": "0600",
"mtime": 1733933711.9187312,
"nlink": 1,
"path": "/etc/cluster/fence_xvm.key",
"pw_name": "root",
"readable": true,
"rgrp": false,
"roth": false,
"rusr": true,
"size": 512,
"uid": 0,
"version": "3811123662",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from cluster nodes fence_xvm.key] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:65
Wednesday 11 December 2024 11:27:42 -0500 (0:00:00.356) 0:00:24.802 ****
ok: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from cluster nodes fence_xvm.key] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:73
Wednesday 11 December 2024 11:27:42 -0500 (0:00:00.529) 0:00:25.331 ****
ok: [managed-node1] => (item=None) => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
ok: [managed-node1 -> {{ item }}] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Send fence_xvm.key to nodes] ******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/distribute-fence-virt-key.yml:17
Wednesday 11 December 2024 11:27:42 -0500 (0:00:00.095) 0:00:25.427 ****
ok: [managed-node1] => {
"changed": false,
"checksum": "fc07c1393356f429049baf24c4f46cce9698df8e",
"dest": "/etc/cluster/fence_xvm.key",
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/etc/cluster/fence_xvm.key",
"secontext": "system_u:object_r:cluster_conf_t:s0",
"size": 512,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Configure SBD] ********************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:69
Wednesday 11 December 2024 11:27:43 -0500 (0:00:00.622) 0:00:26.049 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Configure watchdog kernel module blocklist] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:10
Wednesday 11 December 2024 11:27:43 -0500 (0:00:00.111) 0:00:26.161 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Unload watchdog kernel modules from blocklist] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:22
Wednesday 11 December 2024 11:27:43 -0500 (0:00:00.037) 0:00:26.198 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure watchdog kernel modules] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:30
Wednesday 11 December 2024 11:27:43 -0500 (0:00:00.046) 0:00:26.244 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Load watchdog kernel modules] *****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:40
Wednesday 11 December 2024 11:27:43 -0500 (0:00:00.039) 0:00:26.284 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Probe SBD devices] ****************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:57
Wednesday 11 December 2024 11:27:43 -0500 (0:00:00.037) 0:00:26.322 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Initialize SBD devices] ***********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:70
Wednesday 11 December 2024 11:27:43 -0500 (0:00:00.039) 0:00:26.362 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute SBD config] ************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:88
Wednesday 11 December 2024 11:27:43 -0500 (0:00:00.016) 0:00:26.378 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_sbd_enabled",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Ensure /etc/systemd/system/sbd.service.d directory exists] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:109
Wednesday 11 December 2024 11:27:43 -0500 (0:00:00.042) 0:00:26.420 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_sbd_enabled",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Override start timeout for SBD] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:117
Wednesday 11 December 2024 11:27:43 -0500 (0:00:00.038) 0:00:26.458 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_sbd_enabled",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Reload systemd service files] *****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:131
Wednesday 11 December 2024 11:27:44 -0500 (0:00:00.039) 0:00:26.498 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_sbd_enabled",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Get services status - detect pacemaker] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:135
Wednesday 11 December 2024 11:27:44 -0500 (0:00:00.038) 0:00:26.537 ****
ok: [managed-node1] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"auto-cpufreq.service": {
"name": "auto-cpufreq.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qdevice.service": {
"name": "corosync-qdevice.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"display-manager.service": {
"name": "display-manager.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"fcoe.service": {
"name": "fcoe.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"hv_kvp_daemon.service": {
"name": "hv_kvp_daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ip6tables.service": {
"name": "ip6tables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ipset.service": {
"name": "ipset.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iptables.service": {
"name": "iptables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"iscsi-shutdown.service": {
"name": "iscsi-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iscsi.service": {
"name": "iscsi.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iscsid.service": {
"name": "iscsid.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-activation-early.service": {
"name": "lvm2-activation-early.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"lvm2-activation.service": {
"name": "lvm2-activation.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"modprobe@efi_pstore.service": {
"name": "modprobe@efi_pstore.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"network.service": {
"name": "network.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"ntpd.service": {
"name": "ntpd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ntpdate.service": {
"name": "ntpdate.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"power-profiles-daemon.service": {
"name": "power-profiles-daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rbdmap.service": {
"name": "rbdmap.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"snapd.seeded.service": {
"name": "snapd.seeded.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sntp.service": {
"name": "sntp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen.service": {
"name": "sshd-keygen.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-networkd-wait-online.service": {
"name": "systemd-networkd-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-timesyncd.service": {
"name": "systemd-timesyncd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tlp.service": {
"name": "tlp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"ypbind.service": {
"name": "ypbind.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"yppasswdd.service": {
"name": "yppasswdd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ypserv.service": {
"name": "ypserv.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ypxfrd.service": {
"name": "ypxfrd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Set stonith-watchdog-timeout cluster property in CIB] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:142
Wednesday 11 December 2024 11:27:45 -0500 (0:00:01.568) 0:00:28.105 ****
changed: [managed-node1] => {
"changed": true,
"cmd": [
"pcs",
"--force",
"--",
"property",
"set",
"stonith-watchdog-timeout=0"
],
"delta": "0:00:00.577692",
"end": "2024-12-11 11:27:46.519846",
"rc": 0,
"start": "2024-12-11 11:27:45.942154"
}
TASK [fedora.linux_system_roles.ha_cluster : Correct cib.xml ownership] ********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:155
Wednesday 11 December 2024 11:27:46 -0500 (0:00:00.983) 0:00:29.089 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "not pacemaker_running",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Clean cib.xml.sig] ****************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:167
Wednesday 11 December 2024 11:27:46 -0500 (0:00:00.104) 0:00:29.193 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "not pacemaker_running",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure corosync] ***************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:72
Wednesday 11 December 2024 11:27:46 -0500 (0:00:00.067) 0:00:29.260 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Create a corosync.conf tempfile] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:3
Wednesday 11 December 2024 11:27:46 -0500 (0:00:00.086) 0:00:29.346 ****
changed: [managed-node1] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.utca8fl8_ha_cluster_corosync_conf",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Create a corosync.conf file content] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:16
Wednesday 11 December 2024 11:27:47 -0500 (0:00:00.466) 0:00:29.813 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cluster-setup-pcs-0.10.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Create a corosync.conf file content using pcs-0.10] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cluster-setup-pcs-0.10.yml:3
Wednesday 11 December 2024 11:27:47 -0500 (0:00:00.060) 0:00:29.873 ****
changed: [managed-node1] => {
"changed": true,
"cmd": [
"pcs",
"cluster",
"setup",
"--corosync_conf",
"/tmp/ansible.utca8fl8_ha_cluster_corosync_conf",
"--overwrite",
"--no-cluster-uuid",
"--",
"test-cluster",
"localhost"
],
"delta": "0:00:00.517034",
"end": "2024-12-11 11:27:48.245358",
"rc": 0,
"start": "2024-12-11 11:27:47.728324"
}
STDOUT:
No addresses specified for host 'localhost', using 'localhost'
TASK [fedora.linux_system_roles.ha_cluster : Add qdevice configuration to corosync.conf] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cluster-setup-pcs-0.10.yml:81
Wednesday 11 December 2024 11:27:48 -0500 (0:00:00.900) 0:00:30.774 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__ha_cluster_qdevice_in_use",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch created corosync.conf file] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:19
Wednesday 11 December 2024 11:27:48 -0500 (0:00:00.042) 0:00:30.817 ****
ok: [managed-node1] => {
"changed": false,
"content": "dG90ZW0gewogICAgdmVyc2lvbjogMgogICAgY2x1c3Rlcl9uYW1lOiB0ZXN0LWNsdXN0ZXIKICAgIHRyYW5zcG9ydDoga25ldAogICAgY3J5cHRvX2NpcGhlcjogYWVzMjU2CiAgICBjcnlwdG9faGFzaDogc2hhMjU2Cn0KCm5vZGVsaXN0IHsKICAgIG5vZGUgewogICAgICAgIHJpbmcwX2FkZHI6IGxvY2FsaG9zdAogICAgICAgIG5hbWU6IGxvY2FsaG9zdAogICAgICAgIG5vZGVpZDogMQogICAgfQp9CgpxdW9ydW0gewogICAgcHJvdmlkZXI6IGNvcm9zeW5jX3ZvdGVxdW9ydW0KfQoKbG9nZ2luZyB7CiAgICB0b19sb2dmaWxlOiB5ZXMKICAgIGxvZ2ZpbGU6IC92YXIvbG9nL2NsdXN0ZXIvY29yb3N5bmMubG9nCiAgICB0b19zeXNsb2c6IHllcwogICAgdGltZXN0YW1wOiBvbgp9Cg==",
"encoding": "base64",
"source": "/tmp/ansible.utca8fl8_ha_cluster_corosync_conf"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute corosync.conf file] ****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:26
Wednesday 11 December 2024 11:27:48 -0500 (0:00:00.349) 0:00:31.166 ****
ok: [managed-node1] => {
"changed": false,
"checksum": "05d2ec2a2bfa233bb2b4ace4aae02b42cafc012b",
"dest": "/etc/corosync/corosync.conf",
"gid": 0,
"group": "root",
"mode": "0644",
"owner": "root",
"path": "/etc/corosync/corosync.conf",
"secontext": "system_u:object_r:etc_t:s0",
"size": 388,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Remove a corosync.conf tempfile] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:36
Wednesday 11 December 2024 11:27:49 -0500 (0:00:00.741) 0:00:31.908 ****
changed: [managed-node1] => {
"changed": true,
"path": "/tmp/ansible.utca8fl8_ha_cluster_corosync_conf",
"state": "absent"
}
TASK [fedora.linux_system_roles.ha_cluster : Cluster auth] *********************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:75
Wednesday 11 December 2024 11:27:49 -0500 (0:00:00.403) 0:00:32.311 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-auth.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Check pcs auth status] ************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-auth.yml:3
Wednesday 11 December 2024 11:27:49 -0500 (0:00:00.066) 0:00:32.377 ****
ok: [managed-node1] => {
"changed": false,
"cmd": [
"pcs",
"status",
"pcsd",
"--",
"localhost"
],
"delta": "0:00:00.531546",
"end": "2024-12-11 11:27:50.752558",
"failed_when_result": false,
"rc": 0,
"start": "2024-12-11 11:27:50.221012"
}
STDOUT:
localhost: Online
TASK [fedora.linux_system_roles.ha_cluster : Run pcs auth] *********************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-auth.yml:23
Wednesday 11 December 2024 11:27:50 -0500 (0:00:00.897) 0:00:33.275 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "not ansible_check_mode and __ha_cluster_pcs_auth_status.rc == 2",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute cluster shared keys] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:80
Wednesday 11 December 2024 11:27:50 -0500 (0:00:00.039) 0:00:33.314 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Get corosync authkey] *************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:3
Wednesday 11 December 2024 11:27:50 -0500 (0:00:00.069) 0:00:33.383 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on the controller corosync authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:16
Wednesday 11 December 2024 11:27:50 -0500 (0:00:00.058) 0:00:33.442 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "preshared_key_src is string and preshared_key_src | length > 1",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from the controller corosync authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:21
Wednesday 11 December 2024 11:27:51 -0500 (0:00:00.055) 0:00:33.498 ****
skipping: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from the controller corosync authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:29
Wednesday 11 December 2024 11:27:51 -0500 (0:00:00.051) 0:00:33.549 ****
skipping: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Generate key using OpenSSL corosync authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:44
Wednesday 11 December 2024 11:27:51 -0500 (0:00:00.063) 0:00:33.613 ****
ok: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch generated corosync authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:50
Wednesday 11 December 2024 11:27:51 -0500 (0:00:00.374) 0:00:33.987 ****
ok: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on cluster nodes corosync authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:60
Wednesday 11 December 2024 11:27:51 -0500 (0:00:00.049) 0:00:34.036 ****
ok: [managed-node1] => {
"changed": false,
"stat": {
"atime": 1733933734.5146713,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "binary",
"checksum": "c5293101d291cb165344c5e08b25394fd05dd124",
"ctime": 1733933724.7966807,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 406847629,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/octet-stream",
"mode": "0400",
"mtime": 1733933724.5346808,
"nlink": 1,
"path": "/etc/corosync/authkey",
"pw_name": "root",
"readable": true,
"rgrp": false,
"roth": false,
"rusr": true,
"size": 256,
"uid": 0,
"version": "1298587342",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": false,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from cluster nodes corosync authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:65
Wednesday 11 December 2024 11:27:51 -0500 (0:00:00.356) 0:00:34.393 ****
ok: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from cluster nodes corosync authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:73
Wednesday 11 December 2024 11:27:52 -0500 (0:00:00.356) 0:00:34.749 ****
ok: [managed-node1] => (item=None) => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
ok: [managed-node1 -> {{ item }}] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute corosync authkey] ******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:11
Wednesday 11 December 2024 11:27:52 -0500 (0:00:00.088) 0:00:34.838 ****
ok: [managed-node1] => {
"changed": false,
"checksum": "c5293101d291cb165344c5e08b25394fd05dd124",
"dest": "/etc/corosync/authkey",
"gid": 0,
"group": "root",
"mode": "0400",
"owner": "root",
"path": "/etc/corosync/authkey",
"secontext": "system_u:object_r:etc_t:s0",
"size": 256,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Get pacemaker authkey] ************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:20
Wednesday 11 December 2024 11:27:53 -0500 (0:00:00.645) 0:00:35.483 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on the controller pacemaker authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:16
Wednesday 11 December 2024 11:27:53 -0500 (0:00:00.083) 0:00:35.567 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "preshared_key_src is string and preshared_key_src | length > 1",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from the controller pacemaker authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:21
Wednesday 11 December 2024 11:27:53 -0500 (0:00:00.093) 0:00:35.661 ****
skipping: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from the controller pacemaker authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:29
Wednesday 11 December 2024 11:27:53 -0500 (0:00:00.076) 0:00:35.737 ****
skipping: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Generate key using OpenSSL pacemaker authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:44
Wednesday 11 December 2024 11:27:53 -0500 (0:00:00.066) 0:00:35.803 ****
ok: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch generated pacemaker authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:50
Wednesday 11 December 2024 11:27:53 -0500 (0:00:00.381) 0:00:36.185 ****
ok: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on cluster nodes pacemaker authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:60
Wednesday 11 December 2024 11:27:53 -0500 (0:00:00.052) 0:00:36.238 ****
ok: [managed-node1] => {
"changed": false,
"stat": {
"atime": 1733933864.1563864,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "binary",
"checksum": "4e9eb12a7f75d4c570df4a7b9aed0456f2b1a7f3",
"ctime": 1733933726.7196789,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 189,
"gr_name": "haclient",
"inode": 421527700,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/octet-stream",
"mode": "0400",
"mtime": 1733933726.4426792,
"nlink": 1,
"path": "/etc/pacemaker/authkey",
"pw_name": "hacluster",
"readable": true,
"rgrp": false,
"roth": false,
"rusr": true,
"size": 256,
"uid": 189,
"version": "1336465447",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": false,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from cluster nodes pacemaker authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:65
Wednesday 11 December 2024 11:27:54 -0500 (0:00:00.361) 0:00:36.599 ****
ok: [managed-node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from cluster nodes pacemaker authkey] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:73
Wednesday 11 December 2024 11:27:54 -0500 (0:00:00.433) 0:00:37.033 ****
ok: [managed-node1] => (item=None) => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
ok: [managed-node1 -> {{ item }}] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute pacemaker authkey] *****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:28
Wednesday 11 December 2024 11:27:54 -0500 (0:00:00.064) 0:00:37.098 ****
ok: [managed-node1] => {
"changed": false,
"checksum": "4e9eb12a7f75d4c570df4a7b9aed0456f2b1a7f3",
"dest": "/etc/pacemaker/authkey",
"gid": 189,
"group": "haclient",
"mode": "0400",
"owner": "hacluster",
"path": "/etc/pacemaker/authkey",
"secontext": "system_u:object_r:etc_t:s0",
"size": 256,
"state": "file",
"uid": 189
}
TASK [fedora.linux_system_roles.ha_cluster : Remove qdevice certificates [CLI]] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:37
Wednesday 11 December 2024 11:27:55 -0500 (0:00:00.592) 0:00:37.690 ****
changed: [managed-node1] => {
"changed": true,
"cmd": [
"pcs",
"--",
"qdevice",
"net-client",
"destroy"
],
"delta": "0:00:00.512217",
"end": "2024-12-11 11:27:56.162101",
"rc": 0,
"start": "2024-12-11 11:27:55.649884"
}
TASK [fedora.linux_system_roles.ha_cluster : Remove qdevice certificates [API]] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:46
Wednesday 11 December 2024 11:27:56 -0500 (0:00:00.990) 0:00:38.680 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "'corosync.quorum.device.client.model.net.certificates.rest-api.v2' in __ha_cluster_pcsd_capabilities",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Obtain and distribute qdevice certificates [CLI]] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:54
Wednesday 11 December 2024 11:27:56 -0500 (0:00:00.041) 0:00:38.722 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__ha_cluster_qdevice_in_use",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Obtain and distribute qdevice certificates [API]] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:71
Wednesday 11 December 2024 11:27:56 -0500 (0:00:00.038) 0:00:38.760 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__ha_cluster_qdevice_in_use",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Enable or disable cluster services on boot] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:86
Wednesday 11 December 2024 11:27:56 -0500 (0:00:00.038) 0:00:38.799 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Enable or disable configured cluster services on boot] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:3
Wednesday 11 December 2024 11:27:56 -0500 (0:00:00.070) 0:00:38.869 ****
ok: [managed-node1] => (item=corosync) => {
"ansible_loop_var": "item",
"changed": false,
"enabled": true,
"item": "corosync",
"name": "corosync",
"status": {
"ActiveEnterTimestamp": "Wed 2024-12-11 11:26:34 EST",
"ActiveEnterTimestampMonotonic": "1020658073",
"ActiveExitTimestamp": "Wed 2024-12-11 11:26:33 EST",
"ActiveExitTimestampMonotonic": "1019305096",
"ActiveState": "active",
"After": "systemd-journald.socket sysinit.target network-online.target basic.target system.slice",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-12-11 11:26:34 EST",
"AssertTimestampMonotonic": "1020269394",
"Before": "shutdown.target pacemaker.service multi-user.target corosync-qdevice.service",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-12-11 11:26:34 EST",
"ConditionTimestampMonotonic": "1020269317",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/corosync.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Corosync Cluster Engine",
"DevicePolicy": "auto",
"Documentation": "man:corosync man:corosync.conf man:corosync_overview",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/corosync (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "74530",
"ExecMainStartTimestamp": "Wed 2024-12-11 11:26:34 EST",
"ExecMainStartTimestampMonotonic": "1020270509",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/corosync ; argv[]=/usr/sbin/corosync -f $COROSYNC_OPTIONS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStop": "{ path=/usr/sbin/corosync-cfgtool ; argv[]=/usr/sbin/corosync-cfgtool -H --force ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/corosync.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "corosync.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Wed 2024-12-11 11:26:33 EST",
"InactiveEnterTimestampMonotonic": "1019731276",
"InactiveExitTimestamp": "Wed 2024-12-11 11:26:34 EST",
"InactiveExitTimestampMonotonic": "1020270554",
"InvocationID": "a5c17ed8ed9b415194da8435363913f5",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "74530",
"MemoryAccounting": "yes",
"MemoryCurrent": "143663104",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "corosync.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"RequiredBy": "corosync-qdevice.service pacemaker.service",
"Requires": "sysinit.target network-online.target system.slice",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Wed 2024-12-11 11:26:34 EST",
"StateChangeTimestampMonotonic": "1020658073",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "9",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestamp": "Wed 2024-12-11 11:26:34 EST",
"WatchdogTimestampMonotonic": "1020658050",
"WatchdogUSec": "0"
}
}
skipping: [managed-node1] => (item=corosync-qdevice) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "item != 'corosync-qdevice' or __ha_cluster_qdevice_in_use",
"item": "corosync-qdevice",
"skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=pacemaker) => {
"ansible_loop_var": "item",
"changed": false,
"enabled": true,
"item": "pacemaker",
"name": "pacemaker",
"status": {
"ActiveEnterTimestamp": "Wed 2024-12-11 11:26:35 EST",
"ActiveEnterTimestampMonotonic": "1021582981",
"ActiveExitTimestamp": "Wed 2024-12-11 11:26:31 EST",
"ActiveExitTimestampMonotonic": "1017974492",
"ActiveState": "active",
"After": "dbus.service basic.target sysinit.target rsyslog.service system.slice systemd-journald.socket time-sync.target corosync.service network.target resource-agents-deps.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-12-11 11:26:35 EST",
"AssertTimestampMonotonic": "1021580855",
"Before": "multi-user.target shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-12-11 11:26:35 EST",
"ConditionTimestampMonotonic": "1021580854",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/pacemaker.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Pacemaker High Availability Cluster Manager",
"DevicePolicy": "auto",
"Documentation": "man:pacemakerd https://clusterlabs.org/pacemaker/doc/",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/sbd (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "74793",
"ExecMainStartTimestamp": "Wed 2024-12-11 11:26:35 EST",
"ExecMainStartTimestampMonotonic": "1021582338",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pacemakerd ; argv[]=/usr/sbin/pacemakerd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pacemaker.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pacemaker.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Wed 2024-12-11 11:26:32 EST",
"InactiveEnterTimestampMonotonic": "1018383140",
"InactiveExitTimestamp": "Wed 2024-12-11 11:26:35 EST",
"InactiveExitTimestampMonotonic": "1021582981",
"InvocationID": "85dc9a2036b54c20bfe6901278500d7f",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "process",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "74793",
"MemoryAccounting": "yes",
"MemoryCurrent": "49909760",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pacemaker.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "corosync.service system.slice sysinit.target",
"Restart": "on-failure",
"RestartUSec": "1s",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "no",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "25s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Wed 2024-12-11 11:26:35 EST",
"StateChangeTimestampMonotonic": "1021582981",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "7",
"TasksMax": "infinity",
"TimeoutStartUSec": "1min",
"TimeoutStopUSec": "30min",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "simple",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "dbus.service resource-agents-deps.target",
"WatchdogTimestamp": "Wed 2024-12-11 11:26:35 EST",
"WatchdogTimestampMonotonic": "1021582978",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Get services status - detect SBD] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:16
Wednesday 11 December 2024 11:27:57 -0500 (0:00:01.003) 0:00:39.873 ****
ok: [managed-node1] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"auto-cpufreq.service": {
"name": "auto-cpufreq.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qdevice.service": {
"name": "corosync-qdevice.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"display-manager.service": {
"name": "display-manager.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"fcoe.service": {
"name": "fcoe.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"hv_kvp_daemon.service": {
"name": "hv_kvp_daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ip6tables.service": {
"name": "ip6tables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ipset.service": {
"name": "ipset.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iptables.service": {
"name": "iptables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"iscsi-shutdown.service": {
"name": "iscsi-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iscsi.service": {
"name": "iscsi.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iscsid.service": {
"name": "iscsid.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-activation-early.service": {
"name": "lvm2-activation-early.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"lvm2-activation.service": {
"name": "lvm2-activation.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"modprobe@efi_pstore.service": {
"name": "modprobe@efi_pstore.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"network.service": {
"name": "network.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"ntpd.service": {
"name": "ntpd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ntpdate.service": {
"name": "ntpdate.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"power-profiles-daemon.service": {
"name": "power-profiles-daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rbdmap.service": {
"name": "rbdmap.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"snapd.seeded.service": {
"name": "snapd.seeded.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sntp.service": {
"name": "sntp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen.service": {
"name": "sshd-keygen.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-networkd-wait-online.service": {
"name": "systemd-networkd-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-timesyncd.service": {
"name": "systemd-timesyncd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tlp.service": {
"name": "tlp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"ypbind.service": {
"name": "ypbind.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"yppasswdd.service": {
"name": "yppasswdd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ypserv.service": {
"name": "ypserv.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ypxfrd.service": {
"name": "ypxfrd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Enable or disable SBD] ************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:19
Wednesday 11 December 2024 11:27:59 -0500 (0:00:01.622) 0:00:41.495 ****
ok: [managed-node1] => {
"changed": false,
"enabled": false,
"name": "sbd",
"status": {
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "sysinit.target iscsi.service basic.target systemd-modules-load.service system.slice systemd-journald.socket",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "pacemaker.service shutdown.target dlm.service",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "no",
"CanStop": "no",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Shared-storage based fencing daemon",
"DevicePolicy": "auto",
"Documentation": "man:sbd(8)",
"DropInPaths": "/etc/systemd/system/sbd.service.d/override-timeout.conf",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/sbd (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/sbd ; argv[]=/usr/sbin/sbd $SBD_OPTS -p /run/sbd.pid watch ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStop": "{ path=/usr/bin/kill ; argv[]=/usr/bin/kill -TERM $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/sbd.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "sbd.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "sbd.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PIDFile": "/run/sbd.pid",
"PartOf": "corosync.service",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "yes",
"RefuseManualStop": "yes",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target system.slice",
"Restart": "on-abort",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "forking",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "disabled",
"UtmpMode": "init",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Start the cluster and reload corosync.conf] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:89
Wednesday 11 December 2024 11:27:59 -0500 (0:00:00.519) 0:00:42.014 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Get services status - detect corosync-qdevice] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:21
Wednesday 11 December 2024 11:27:59 -0500 (0:00:00.108) 0:00:42.123 ****
ok: [managed-node1] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"auto-cpufreq.service": {
"name": "auto-cpufreq.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qdevice.service": {
"name": "corosync-qdevice.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"display-manager.service": {
"name": "display-manager.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"fcoe.service": {
"name": "fcoe.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"hv_kvp_daemon.service": {
"name": "hv_kvp_daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ip6tables.service": {
"name": "ip6tables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ipset.service": {
"name": "ipset.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iptables.service": {
"name": "iptables.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"iscsi-shutdown.service": {
"name": "iscsi-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iscsi.service": {
"name": "iscsi.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"iscsid.service": {
"name": "iscsid.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-activation-early.service": {
"name": "lvm2-activation-early.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"lvm2-activation.service": {
"name": "lvm2-activation.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"modprobe@efi_pstore.service": {
"name": "modprobe@efi_pstore.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"network.service": {
"name": "network.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"ntpd.service": {
"name": "ntpd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ntpdate.service": {
"name": "ntpdate.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"power-profiles-daemon.service": {
"name": "power-profiles-daemon.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rbdmap.service": {
"name": "rbdmap.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"snapd.seeded.service": {
"name": "snapd.seeded.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sntp.service": {
"name": "sntp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen.service": {
"name": "sshd-keygen.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "inactive"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-networkd-wait-online.service": {
"name": "systemd-networkd-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-timesyncd.service": {
"name": "systemd-timesyncd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tlp.service": {
"name": "tlp.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "active"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "active"
},
"ypbind.service": {
"name": "ypbind.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"yppasswdd.service": {
"name": "yppasswdd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ypserv.service": {
"name": "ypserv.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
},
"ypxfrd.service": {
"name": "ypxfrd.service",
"source": "systemd",
"state": "stopped",
"status": "not-found"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Stop cluster daemons to reload configuration] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:24
Wednesday 11 December 2024 11:28:01 -0500 (0:00:01.633) 0:00:43.756 ****
skipping: [managed-node1] => (item=pacemaker) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__ha_cluster_distribute_corosync_conf.changed or __ha_cluster_distribute_corosync_authkey.changed or __ha_cluster_distribute_pacemaker_authkey.changed or (__ha_cluster_sbd_service_enable_disable.changed | d(false)) or (__ha_cluster_distribute_sbd_config.changed | d(false)) or (__ha_cluster_qdevice_certs.changed | d(false))\n",
"item": "pacemaker",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=corosync-qdevice) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__ha_cluster_distribute_corosync_conf.changed or __ha_cluster_distribute_corosync_authkey.changed or __ha_cluster_distribute_pacemaker_authkey.changed or (__ha_cluster_sbd_service_enable_disable.changed | d(false)) or (__ha_cluster_distribute_sbd_config.changed | d(false)) or (__ha_cluster_qdevice_certs.changed | d(false))\n",
"item": "corosync-qdevice",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=corosync) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "__ha_cluster_distribute_corosync_conf.changed or __ha_cluster_distribute_corosync_authkey.changed or __ha_cluster_distribute_pacemaker_authkey.changed or (__ha_cluster_sbd_service_enable_disable.changed | d(false)) or (__ha_cluster_distribute_sbd_config.changed | d(false)) or (__ha_cluster_qdevice_certs.changed | d(false))\n",
"item": "corosync",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => {
"changed": false
}
MSG:
All items skipped
TASK [fedora.linux_system_roles.ha_cluster : Start corosync] *******************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:47
Wednesday 11 December 2024 11:28:01 -0500 (0:00:00.096) 0:00:43.853 ****
ok: [managed-node1] => {
"changed": false,
"name": "corosync",
"state": "started",
"status": {
"ActiveEnterTimestamp": "Wed 2024-12-11 11:26:34 EST",
"ActiveEnterTimestampMonotonic": "1020658073",
"ActiveExitTimestamp": "Wed 2024-12-11 11:26:33 EST",
"ActiveExitTimestampMonotonic": "1019305096",
"ActiveState": "active",
"After": "systemd-journald.socket sysinit.target network-online.target basic.target system.slice",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-12-11 11:26:34 EST",
"AssertTimestampMonotonic": "1020269394",
"Before": "shutdown.target pacemaker.service multi-user.target corosync-qdevice.service",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-12-11 11:26:34 EST",
"ConditionTimestampMonotonic": "1020269317",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/corosync.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Corosync Cluster Engine",
"DevicePolicy": "auto",
"Documentation": "man:corosync man:corosync.conf man:corosync_overview",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/corosync (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "74530",
"ExecMainStartTimestamp": "Wed 2024-12-11 11:26:34 EST",
"ExecMainStartTimestampMonotonic": "1020270509",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/corosync ; argv[]=/usr/sbin/corosync -f $COROSYNC_OPTIONS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStop": "{ path=/usr/sbin/corosync-cfgtool ; argv[]=/usr/sbin/corosync-cfgtool -H --force ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/corosync.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "corosync.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Wed 2024-12-11 11:26:33 EST",
"InactiveEnterTimestampMonotonic": "1019731276",
"InactiveExitTimestamp": "Wed 2024-12-11 11:26:34 EST",
"InactiveExitTimestampMonotonic": "1020270554",
"InvocationID": "a5c17ed8ed9b415194da8435363913f5",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "74530",
"MemoryAccounting": "yes",
"MemoryCurrent": "143663104",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "corosync.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"RequiredBy": "corosync-qdevice.service pacemaker.service",
"Requires": "sysinit.target network-online.target system.slice",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Wed 2024-12-11 11:26:34 EST",
"StateChangeTimestampMonotonic": "1020658073",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "9",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestamp": "Wed 2024-12-11 11:26:34 EST",
"WatchdogTimestampMonotonic": "1020658050",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Reload corosync configuration] ****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:56
Wednesday 11 December 2024 11:28:02 -0500 (0:00:00.616) 0:00:44.469 ****
ok: [managed-node1] => {
"changed": false,
"cmd": [
"corosync-cfgtool",
"-R"
],
"delta": "0:00:00.011807",
"end": "2024-12-11 11:28:02.313315",
"rc": 0,
"start": "2024-12-11 11:28:02.301508"
}
STDOUT:
Reloading corosync.conf...
Done
TASK [fedora.linux_system_roles.ha_cluster : Start corosync-qdevice] ***********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:63
Wednesday 11 December 2024 11:28:02 -0500 (0:00:00.360) 0:00:44.829 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "__ha_cluster_qdevice_in_use",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Start pacemaker] ******************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:69
Wednesday 11 December 2024 11:28:02 -0500 (0:00:00.039) 0:00:44.868 ****
ok: [managed-node1] => {
"changed": false,
"name": "pacemaker",
"state": "started",
"status": {
"ActiveEnterTimestamp": "Wed 2024-12-11 11:26:35 EST",
"ActiveEnterTimestampMonotonic": "1021582981",
"ActiveExitTimestamp": "Wed 2024-12-11 11:26:31 EST",
"ActiveExitTimestampMonotonic": "1017974492",
"ActiveState": "active",
"After": "dbus.service basic.target sysinit.target rsyslog.service system.slice systemd-journald.socket time-sync.target corosync.service network.target resource-agents-deps.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Wed 2024-12-11 11:26:35 EST",
"AssertTimestampMonotonic": "1021580855",
"Before": "multi-user.target shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Wed 2024-12-11 11:26:35 EST",
"ConditionTimestampMonotonic": "1021580854",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/pacemaker.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Pacemaker High Availability Cluster Manager",
"DevicePolicy": "auto",
"Documentation": "man:pacemakerd https://clusterlabs.org/pacemaker/doc/",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/sbd (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "74793",
"ExecMainStartTimestamp": "Wed 2024-12-11 11:26:35 EST",
"ExecMainStartTimestampMonotonic": "1021582338",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pacemakerd ; argv[]=/usr/sbin/pacemakerd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pacemaker.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pacemaker.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Wed 2024-12-11 11:26:32 EST",
"InactiveEnterTimestampMonotonic": "1018383140",
"InactiveExitTimestamp": "Wed 2024-12-11 11:26:35 EST",
"InactiveExitTimestampMonotonic": "1021582981",
"InvocationID": "85dc9a2036b54c20bfe6901278500d7f",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "process",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "74793",
"MemoryAccounting": "yes",
"MemoryCurrent": "52006912",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pacemaker.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "corosync.service system.slice sysinit.target",
"Restart": "on-failure",
"RestartUSec": "1s",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "no",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "25s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Wed 2024-12-11 11:26:35 EST",
"StateChangeTimestampMonotonic": "1021582981",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "7",
"TasksMax": "infinity",
"TimeoutStartUSec": "1min",
"TimeoutStopUSec": "30min",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "simple",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "dbus.service resource-agents-deps.target",
"WatchdogTimestamp": "Wed 2024-12-11 11:26:35 EST",
"WatchdogTimestampMonotonic": "1021582978",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Wait for the cluster to fully start and form membership] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:75
Wednesday 11 December 2024 11:28:02 -0500 (0:00:00.533) 0:00:45.402 ****
ok: [managed-node1] => {
"changed": false,
"cmd": [
"pcs",
"cluster",
"start",
"--all",
"--wait"
],
"delta": "0:00:05.279109",
"end": "2024-12-11 11:28:08.535317",
"rc": 0,
"start": "2024-12-11 11:28:03.256208"
}
STDOUT:
localhost: Starting Cluster...
Waiting for node(s) to start...
localhost: Started
TASK [fedora.linux_system_roles.ha_cluster : List pacemaker nodes] *************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:89
Wednesday 11 December 2024 11:28:08 -0500 (0:00:05.653) 0:00:51.055 ****
ok: [managed-node1] => {
"changed": false,
"cmd": "set -euo pipefail; crm_mon -X | xmllint --xpath '/crm_mon/nodes/node/@name' - | sed -E 's/\\s*name=\"([^\"]+)\"\\s*/\\1\\n/g'\n",
"delta": "0:00:00.017784",
"end": "2024-12-11 11:28:08.915467",
"rc": 0,
"start": "2024-12-11 11:28:08.897683"
}
STDOUT:
localhost
TASK [fedora.linux_system_roles.ha_cluster : Purge removed nodes from pacemaker's cache] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:111
Wednesday 11 December 2024 11:28:08 -0500 (0:00:00.375) 0:00:51.430 ****
skipping: [managed-node1] => (item=localhost) => {
"ansible_loop_var": "item",
"changed": false,
"false_condition": "item not in __ha_cluster_all_node_names",
"item": "localhost",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => {
"changed": false
}
MSG:
All items skipped
TASK [fedora.linux_system_roles.ha_cluster : Create and push CIB] **************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:92
Wednesday 11 December 2024 11:28:09 -0500 (0:00:00.053) 0:00:51.484 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Create a tempfile for original CIB] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:10
Wednesday 11 December 2024 11:28:09 -0500 (0:00:00.157) 0:00:51.641 ****
changed: [managed-node1] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.j58oeqir_ha_cluster_original_cib_xml",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Create a tempfile for new CIB] ****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:18
Wednesday 11 December 2024 11:28:09 -0500 (0:00:00.360) 0:00:52.001 ****
changed: [managed-node1] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.4vvd3og8_ha_cluster_cib_xml",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch CIB configuration] **********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:26
Wednesday 11 December 2024 11:28:09 -0500 (0:00:00.378) 0:00:52.379 ****
ok: [managed-node1] => {
"changed": false,
"cmd": [
"cibadmin",
"--query"
],
"delta": "0:00:00.010168",
"end": "2024-12-11 11:28:10.218559",
"rc": 0,
"start": "2024-12-11 11:28:10.208391"
}
STDOUT:
TASK [fedora.linux_system_roles.ha_cluster : Write CIB configuration] **********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:33
Wednesday 11 December 2024 11:28:10 -0500 (0:00:00.352) 0:00:52.732 ****
changed: [managed-node1] => (item=/tmp/ansible.4vvd3og8_ha_cluster_cib_xml) => {
"ansible_loop_var": "item",
"changed": true,
"checksum": "dc4f050a21bbfb30aabb4a8d5c1d66e99a4d0d5a",
"dest": "/tmp/ansible.4vvd3og8_ha_cluster_cib_xml",
"gid": 0,
"group": "root",
"item": "/tmp/ansible.4vvd3og8_ha_cluster_cib_xml",
"md5sum": "fc41bd35933cf480ef9b5eb5470fda05",
"mode": "0600",
"owner": "root",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 2454,
"src": "/root/.ansible/tmp/ansible-tmp-1733934490.3256793-37184-144029677271015/source",
"state": "file",
"uid": 0
}
changed: [managed-node1] => (item=/tmp/ansible.j58oeqir_ha_cluster_original_cib_xml) => {
"ansible_loop_var": "item",
"changed": true,
"checksum": "dc4f050a21bbfb30aabb4a8d5c1d66e99a4d0d5a",
"dest": "/tmp/ansible.j58oeqir_ha_cluster_original_cib_xml",
"gid": 0,
"group": "root",
"item": "/tmp/ansible.j58oeqir_ha_cluster_original_cib_xml",
"md5sum": "fc41bd35933cf480ef9b5eb5470fda05",
"mode": "0600",
"owner": "root",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 2454,
"src": "/root/.ansible/tmp/ansible-tmp-1733934491.1237025-37184-55201489672516/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Purge new CIB configuration] ******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:51
Wednesday 11 December 2024 11:28:11 -0500 (0:00:01.441) 0:00:54.174 ****
changed: [managed-node1] => {
"changed": true,
"cmd": [
"cibadmin",
"--force",
"--delete-all",
"--xpath",
"/cib/configuration/*[not(\n self::crm_config or\n self::nodes or\n self::resources or\n self::constraints\n)] | /cib/configuration/*[self::resources or self::constraints]/* | /cib/configuration/nodes/*/* | /cib/configuration/crm_config//nvpair[not(\n @name=\"cluster-infrastructure\" or\n @name=\"cluster-name\" or\n @name=\"dc-version\" or\n @name=\"have-watchdog\" or\n @name=\"last-lrm-refresh\" or\n @name=\"stonith-watchdog-timeout\"\n)]"
],
"delta": "0:00:00.022378",
"end": "2024-12-11 11:28:12.023971",
"rc": 0,
"start": "2024-12-11 11:28:12.001593"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster properties] *****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:85
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.365) 0:00:54.540 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_cluster_properties[0].attrs | d([])",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure node attributes] ********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:95
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.043) 0:00:54.583 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure node utilization] *******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:102
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.044) 0:00:54.628 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure resource defaults] ******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:115
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.039) 0:00:54.668 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure resource operation defaults] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:127
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.038) 0:00:54.706 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Extract primitive to bundle mapping] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:137
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.044) 0:00:54.751 ****
ok: [managed-node1] => {
"ansible_facts": {
"__ha_cluster_primitive_bundle_map": {}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster bundle resources] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:141
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.055) 0:00:54.807 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster resources] ******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:147
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.026) 0:00:54.833 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster resource groups] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:154
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.028) 0:00:54.861 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster resource clones] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:160
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.028) 0:00:54.890 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure stonith levels] *********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:167
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.030) 0:00:54.920 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure resource location constraints] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:175
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.034) 0:00:54.955 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure resource colocation constraints] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:182
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.046) 0:00:55.001 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure resource set colocation constraints] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:190
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.028) 0:00:55.029 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure resource order constraints] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:200
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.028) 0:00:55.058 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure resource set order constraints] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:208
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.027) 0:00:55.086 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure resource ticket constraints] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:218
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.028) 0:00:55.114 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure resource set ticket constraints] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:226
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.026) 0:00:55.141 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure acls] *******************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:236
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.030) 0:00:55.172 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cib-acls.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Configure ACL roles] **************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cib-acls.yml:3
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.081) 0:00:55.254 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure ACL users] **************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cib-acls.yml:28
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.043) 0:00:55.297 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure ACL groups] *************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cib-acls.yml:44
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.043) 0:00:55.340 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure alerts] *****************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:241
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.043) 0:00:55.384 ****
skipping: [managed-node1] => {
"changed": false,
"skipped_reason": "No items in the list"
}
TASK [fedora.linux_system_roles.ha_cluster : Create a tempfile for CIB diff] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:249
Wednesday 11 December 2024 11:28:12 -0500 (0:00:00.043) 0:00:55.427 ****
changed: [managed-node1] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.b637q8_b_ha_cluster_cib_diff",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Compare new and original CIB] *****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:257
Wednesday 11 December 2024 11:28:13 -0500 (0:00:00.357) 0:00:55.785 ****
changed: [managed-node1] => {
"changed": true,
"cmd": [
"crm_diff",
"--no-version",
"--original",
"/tmp/ansible.j58oeqir_ha_cluster_original_cib_xml",
"--new",
"/tmp/ansible.4vvd3og8_ha_cluster_cib_xml"
],
"delta": "0:00:00.005870",
"end": "2024-12-11 11:28:13.627065",
"failed_when_result": false,
"rc": 1,
"start": "2024-12-11 11:28:13.621195"
}
STDOUT:
MSG:
non-zero return code
TASK [fedora.linux_system_roles.ha_cluster : Write CIB diff to its tempfile] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:270
Wednesday 11 December 2024 11:28:13 -0500 (0:00:00.359) 0:00:56.144 ****
changed: [managed-node1] => {
"changed": true,
"checksum": "08b2c29ac36b69932c1a533e8a6a3049e43c1392",
"dest": "/tmp/ansible.b637q8_b_ha_cluster_cib_diff",
"gid": 0,
"group": "root",
"md5sum": "5646fc09b0507b078651c052e7815bd6",
"mode": "0600",
"owner": "root",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 89,
"src": "/root/.ansible/tmp/ansible-tmp-1733934493.7353494-37317-220451844067664/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Push CIB diff to the cluster if it has any changes] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:281
Wednesday 11 December 2024 11:28:14 -0500 (0:00:00.643) 0:00:56.787 ****
changed: [managed-node1] => {
"changed": true,
"cmd": [
"cibadmin",
"--verbose",
"--patch",
"--xml-file",
"/tmp/ansible.b637q8_b_ha_cluster_cib_diff"
],
"delta": "0:00:00.014986",
"end": "2024-12-11 11:28:14.672110",
"rc": 0,
"start": "2024-12-11 11:28:14.657124"
}
TASK [fedora.linux_system_roles.ha_cluster : Remove CIB tempfiles] *************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:290
Wednesday 11 December 2024 11:28:14 -0500 (0:00:00.403) 0:00:57.191 ****
changed: [managed-node1] => (item={'changed': True, 'path': '/tmp/ansible.4vvd3og8_ha_cluster_cib_xml', 'uid': 0, 'gid': 0, 'owner': 'root', 'group': 'root', 'mode': '0600', 'state': 'file', 'secontext': 'unconfined_u:object_r:user_tmp_t:s0', 'size': 0, 'failed': False}) => {
"ansible_loop_var": "item",
"changed": true,
"item": {
"changed": true,
"failed": false,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.4vvd3og8_ha_cluster_cib_xml",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
},
"path": "/tmp/ansible.4vvd3og8_ha_cluster_cib_xml",
"state": "absent"
}
changed: [managed-node1] => (item={'changed': True, 'path': '/tmp/ansible.j58oeqir_ha_cluster_original_cib_xml', 'uid': 0, 'gid': 0, 'owner': 'root', 'group': 'root', 'mode': '0600', 'state': 'file', 'secontext': 'unconfined_u:object_r:user_tmp_t:s0', 'size': 0, 'failed': False}) => {
"ansible_loop_var": "item",
"changed": true,
"item": {
"changed": true,
"failed": false,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.j58oeqir_ha_cluster_original_cib_xml",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
},
"path": "/tmp/ansible.j58oeqir_ha_cluster_original_cib_xml",
"state": "absent"
}
changed: [managed-node1] => (item={'changed': True, 'path': '/tmp/ansible.b637q8_b_ha_cluster_cib_diff', 'uid': 0, 'gid': 0, 'owner': 'root', 'group': 'root', 'mode': '0600', 'state': 'file', 'secontext': 'unconfined_u:object_r:user_tmp_t:s0', 'size': 0, 'failed': False}) => {
"ansible_loop_var": "item",
"changed": true,
"item": {
"changed": true,
"failed": false,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.b637q8_b_ha_cluster_cib_diff",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
},
"path": "/tmp/ansible.b637q8_b_ha_cluster_cib_diff",
"state": "absent"
}
TASK [fedora.linux_system_roles.ha_cluster : Remove cluster configuration] *****
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:101
Wednesday 11 December 2024 11:28:15 -0500 (0:00:01.026) 0:00:58.218 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "not ha_cluster_cluster_present",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Remove fence-virt authkey] ********
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:104
Wednesday 11 December 2024 11:28:15 -0500 (0:00:00.049) 0:00:58.268 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "not ha_cluster_cluster_present",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure qnetd] ******************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:109
Wednesday 11 December 2024 11:28:15 -0500 (0:00:00.037) 0:00:58.305 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml for managed-node1
TASK [fedora.linux_system_roles.ha_cluster : Remove qnetd configuration] *******
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:3
Wednesday 11 December 2024 11:28:15 -0500 (0:00:00.082) 0:00:58.388 ****
changed: [managed-node1] => {
"changed": true,
"cmd": [
"pcs",
"--force",
"--",
"qdevice",
"destroy",
"net"
],
"delta": "0:00:01.034512",
"end": "2024-12-11 11:28:17.439826",
"rc": 0,
"start": "2024-12-11 11:28:16.405314"
}
STDOUT:
Stopping quorum device...
quorum device stopped
quorum device disabled
Quorum device 'net' configuration files removed
TASK [fedora.linux_system_roles.ha_cluster : Setup qnetd] **********************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:16
Wednesday 11 December 2024 11:28:17 -0500 (0:00:01.588) 0:00:59.977 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_qnetd.present | d(false)",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Enable or disable qnetd service on boot] ***
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:26
Wednesday 11 December 2024 11:28:17 -0500 (0:00:00.050) 0:01:00.027 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_qnetd.present | d(false)",
"skip_reason": "Conditional result was False"
}
TASK [Fetch cluster properties configuration from the cluster] *****************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cib_properties_empty.yml:33
Wednesday 11 December 2024 11:28:17 -0500 (0:00:00.058) 0:01:00.086 ****
ok: [managed-node1] => {
"changed": false,
"cmd": [
"pcs",
"property"
],
"delta": "0:00:00.637529",
"end": "2024-12-11 11:28:18.632887",
"rc": 0,
"start": "2024-12-11 11:28:17.995358"
}
STDOUT:
Cluster Properties:
cluster-infrastructure: corosync
cluster-name: test-cluster
dc-version: 2.1.7-5.el8-0f7f88312
have-watchdog: false
stonith-watchdog-timeout: 0
TASK [Debug cluster properties configuration] **********************************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cib_properties_empty.yml:39
Wednesday 11 December 2024 11:28:18 -0500 (0:00:01.107) 0:01:01.193 ****
ok: [managed-node1] => {
"__test_pcs_property_config": {
"changed": false,
"cmd": [
"pcs",
"property"
],
"delta": "0:00:00.637529",
"end": "2024-12-11 11:28:18.632887",
"failed": false,
"msg": "",
"rc": 0,
"start": "2024-12-11 11:28:17.995358",
"stderr": "",
"stderr_lines": [],
"stdout": "Cluster Properties:\n cluster-infrastructure: corosync\n cluster-name: test-cluster\n dc-version: 2.1.7-5.el8-0f7f88312\n have-watchdog: false\n stonith-watchdog-timeout: 0",
"stdout_lines": [
"Cluster Properties:",
" cluster-infrastructure: corosync",
" cluster-name: test-cluster",
" dc-version: 2.1.7-5.el8-0f7f88312",
" have-watchdog: false",
" stonith-watchdog-timeout: 0"
]
}
}
TASK [Check firewall and selinux state] ****************************************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cib_properties_empty.yml:43
Wednesday 11 December 2024 11:28:18 -0500 (0:00:00.038) 0:01:01.232 ****
included: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml for managed-node1
TASK [Check firewall service status] *******************************************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:6
Wednesday 11 December 2024 11:28:18 -0500 (0:00:00.062) 0:01:01.295 ****
ok: [managed-node1] => {
"changed": false,
"cmd": [
"firewall-cmd",
"--list-services"
],
"delta": "0:00:00.402854",
"end": "2024-12-11 11:28:19.667670",
"failed_when_result": false,
"rc": 0,
"start": "2024-12-11 11:28:19.264816"
}
STDOUT:
cockpit dhcpv6-client high-availability ssh
TASK [Check firewall port status] **********************************************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:12
Wednesday 11 December 2024 11:28:19 -0500 (0:00:00.915) 0:01:02.211 ****
ok: [managed-node1] => {
"changed": false,
"cmd": [
"firewall-cmd",
"--list-ports"
],
"delta": "0:00:00.399537",
"end": "2024-12-11 11:28:20.578404",
"failed_when_result": false,
"rc": 0,
"start": "2024-12-11 11:28:20.178867"
}
STDOUT:
1229/tcp
TASK [Get associated selinux ports] ********************************************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:25
Wednesday 11 December 2024 11:28:20 -0500 (0:00:00.918) 0:01:03.129 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_manage_selinux | bool",
"skip_reason": "Conditional result was False"
}
TASK [Check associated selinux ports] ******************************************
task path: /tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:33
Wednesday 11 December 2024 11:28:20 -0500 (0:00:00.064) 0:01:03.194 ****
skipping: [managed-node1] => {
"changed": false,
"false_condition": "ha_cluster_manage_selinux | bool",
"skip_reason": "Conditional result was False"
}
PLAY RECAP *********************************************************************
managed-node1 : ok=110 changed=17 unreachable=0 failed=0 skipped=99 rescued=0 ignored=0
TASKS RECAP ********************************************************************
Wednesday 11 December 2024 11:28:20 -0500 (0:00:00.057) 0:01:03.251 ****
===============================================================================
fedora.linux_system_roles.ha_cluster : Wait for the cluster to fully start and form membership --- 5.65s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:75
fedora.linux_system_roles.ha_cluster : Install role essential packages --- 2.98s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:11
fedora.linux_system_roles.firewall : Install firewalld ------------------ 2.94s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31
fedora.linux_system_roles.ha_cluster : Install cluster packages --------- 2.87s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:58
fedora.linux_system_roles.ha_cluster : Populate service facts ----------- 1.67s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:3
fedora.linux_system_roles.ha_cluster : Get services status - detect corosync-qdevice --- 1.63s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:21
fedora.linux_system_roles.ha_cluster : Get services status - detect SBD --- 1.62s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:16
fedora.linux_system_roles.ha_cluster : Remove qnetd configuration ------- 1.59s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:3
fedora.linux_system_roles.ha_cluster : Get services status - detect pacemaker --- 1.57s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:135
fedora.linux_system_roles.ha_cluster : Write CIB configuration ---------- 1.44s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:33
fedora.linux_system_roles.firewall : Configure firewall ----------------- 1.19s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71
fedora.linux_system_roles.ha_cluster : Start pcsd with updated config files and configure it to start on boot --- 1.18s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:88
Gathering Facts --------------------------------------------------------- 1.15s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cib_properties_empty.yml:9
Fetch cluster properties configuration from the cluster ----------------- 1.11s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cib_properties_empty.yml:33
fedora.linux_system_roles.ha_cluster : Remove CIB tempfiles ------------- 1.03s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:290
fedora.linux_system_roles.ha_cluster : Enable or disable configured cluster services on boot --- 1.00s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:3
fedora.linux_system_roles.ha_cluster : Remove qdevice certificates [CLI] --- 0.99s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:37
fedora.linux_system_roles.ha_cluster : Set stonith-watchdog-timeout cluster property in CIB --- 0.98s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:142
Check firewall port status ---------------------------------------------- 0.92s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:12
Check firewall service status ------------------------------------------- 0.92s
/tmp/collections-SQL/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:6