|
@@ -1,6 +1,7 @@
|
|
---
|
|
---
|
|
# we want to ignore if this fails because it is possible to fail on the first node when installing -- should we not do this if we're skipping health checks? -- is this required for cluster sanity?
|
|
# we want to ignore if this fails because it is possible to fail on the first node when installing -- should we not do this if we're skipping health checks? -- is this required for cluster sanity?
|
|
-- name: "Disable shard balancing for logging-{{ _cluster_component }} cluster"
|
|
|
|
|
|
+- when: not _skip_healthcheck | bool
|
|
|
|
+ name: "Disable shard balancing for logging-{{ _cluster_component }} cluster"
|
|
command: >
|
|
command: >
|
|
curl -s -k
|
|
curl -s -k
|
|
--cert {{ _logging_handler_tempdir.stdout }}/admin-cert
|
|
--cert {{ _logging_handler_tempdir.stdout }}/admin-cert
|
|
@@ -20,8 +21,8 @@
|
|
rollout latest {{ _es_node }}
|
|
rollout latest {{ _es_node }}
|
|
-n {{ openshift_logging_elasticsearch_namespace }}
|
|
-n {{ openshift_logging_elasticsearch_namespace }}
|
|
|
|
|
|
-# always wait for this to scale up
|
|
|
|
-- name: "Waiting for {{ _es_node }} to finish scaling up"
|
|
|
|
|
|
+- when: not _skip_healthcheck | bool
|
|
|
|
+ name: "Waiting for {{ _es_node }} to finish scaling up"
|
|
oc_obj:
|
|
oc_obj:
|
|
state: list
|
|
state: list
|
|
name: "{{ _es_node }}"
|
|
name: "{{ _es_node }}"
|
|
@@ -47,7 +48,8 @@
|
|
installer_phase_logging:
|
|
installer_phase_logging:
|
|
message: "Node {{ _es_node}} in cluster logging-{{ _cluster_component }} was unable to rollout. Please see documentation regarding recovering during a {{ 'full' if full_restart_cluster | bool else 'rolling'}} cluster restart."
|
|
message: "Node {{ _es_node}} in cluster logging-{{ _cluster_component }} was unable to rollout. Please see documentation regarding recovering during a {{ 'full' if full_restart_cluster | bool else 'rolling'}} cluster restart."
|
|
|
|
|
|
-- name: "Enable shard balancing for logging-{{ _cluster_component }} cluster"
|
|
|
|
|
|
+- when: not _skip_healthcheck | bool
|
|
|
|
+ name: "Enable shard balancing for logging-{{ _cluster_component }} cluster"
|
|
command: >
|
|
command: >
|
|
curl -s -k
|
|
curl -s -k
|
|
--cert {{ _logging_handler_tempdir.stdout }}/admin-cert
|
|
--cert {{ _logging_handler_tempdir.stdout }}/admin-cert
|
|
@@ -72,19 +74,19 @@
|
|
--cert {{ _logging_handler_tempdir.stdout }}/admin-cert
|
|
--cert {{ _logging_handler_tempdir.stdout }}/admin-cert
|
|
--key {{ _logging_handler_tempdir.stdout }}/admin-key
|
|
--key {{ _logging_handler_tempdir.stdout }}/admin-key
|
|
https://logging-{{ _cluster_component }}.{{ openshift_logging_elasticsearch_namespace }}.svc:9200/_cluster/health?pretty
|
|
https://logging-{{ _cluster_component }}.{{ openshift_logging_elasticsearch_namespace }}.svc:9200/_cluster/health?pretty
|
|
- register: _pod_status
|
|
|
|
|
|
+ register: _cluster_status
|
|
until:
|
|
until:
|
|
- - "_pod_status.stdout != ''"
|
|
|
|
- - (_pod_status.stdout | from_json)['status'] in ['green']
|
|
|
|
|
|
+ - "_cluster_status.stdout != ''"
|
|
|
|
+ - (_cluster_status.stdout | from_json)['status'] in ['green']
|
|
retries: "{{ __elasticsearch_ready_retries }}"
|
|
retries: "{{ __elasticsearch_ready_retries }}"
|
|
delay: 30
|
|
delay: 30
|
|
changed_when: false
|
|
changed_when: false
|
|
failed_when: false
|
|
failed_when: false
|
|
|
|
|
|
-# evaluate RC for _pod_status
|
|
|
|
|
|
+# evaluate RC for _cluster_status
|
|
- when:
|
|
- when:
|
|
- - _pod_status.failed is defined
|
|
|
|
- - _pod_status.failed
|
|
|
|
|
|
+ - _cluster_status.failed is defined
|
|
|
|
+ - _cluster_status.failed
|
|
run_once: true
|
|
run_once: true
|
|
set_stats:
|
|
set_stats:
|
|
data:
|
|
data:
|
|
@@ -93,5 +95,5 @@
|
|
|
|
|
|
- name: Evaluating cluster health
|
|
- name: Evaluating cluster health
|
|
assert:
|
|
assert:
|
|
- that: _pod_status.failed is undefined or not _pod_status.failed
|
|
|
|
|
|
+ that: _cluster_status.failed is undefined or not _cluster_status.failed
|
|
msg: "Cluster logging-{{ _cluster_component }} was unable to recover to a green state. Please see documentation regarding recovering during a {{ 'full' if full_restart_cluster | bool else 'rolling'}} cluster restart."
|
|
msg: "Cluster logging-{{ _cluster_component }} was unable to recover to a green state. Please see documentation regarding recovering during a {{ 'full' if full_restart_cluster | bool else 'rolling'}} cluster restart."
|