restart_cluster.yml 3.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384
  1. ---
  2. ## get all pods for the cluster
  3. - command: >
  4. {{ openshift_client_binary }}
  5. --config={{ openshift.common.config_base }}/master/admin.kubeconfig
  6. get pod
  7. -l component={{ _cluster_component }},provider=openshift
  8. -n {{ openshift_logging_elasticsearch_namespace }}
  9. -o jsonpath={.items[?(@.status.phase==\"Running\")].metadata.name}
  10. register: _cluster_pods
  11. # make a temp dir for admin certs
  12. - command: mktemp -d /tmp/openshift-logging-ansible-XXXXXX
  13. register: _logging_handler_tempdir
  14. changed_when: False
  15. check_mode: no
  16. - name: Exporting secrets to use communicating with the ES cluster
  17. command: >
  18. {{ openshift_client_binary }}
  19. --config={{ openshift.common.config_base }}/master/admin.kubeconfig
  20. extract secret/logging-elasticsearch
  21. -n {{ openshift_logging_elasticsearch_namespace }}
  22. --keys=admin-cert --keys=admin-key
  23. --to={{ _logging_handler_tempdir.stdout }}
  24. ### Check for cluster state before making changes -- if its red, yellow or missing nodes then we don't want to continue
  25. - name: "Checking current health for {{ _es_node }} cluster"
  26. command: >
  27. curl -s -k
  28. --cert {{ _logging_handler_tempdir.stdout }}/admin-cert
  29. --key {{ _logging_handler_tempdir.stdout }}/admin-key
  30. https://logging-{{ _cluster_component }}.{{ openshift_logging_elasticsearch_namespace }}.svc:9200/_cluster/health?pretty
  31. register: _pod_status
  32. when:
  33. - _cluster_pods.stdout
  34. - _cluster_pods.stdout.split(' ') | count > 0
  35. - when:
  36. - _pod_status.stdout is defined
  37. - (_pod_status.stdout | from_json)['status'] in ['yellow', 'red'] or (_pod_status.stdout | from_json)['number_of_nodes'] != _cluster_pods.stdout.split(' ') | count
  38. block:
  39. - name: Set Logging message to manually restart
  40. run_once: true
  41. set_stats:
  42. data:
  43. installer_phase_logging:
  44. message: "Cluster logging-{{ _cluster_component }} was not in an optimal state and will not be automatically restarted. Please see documentation regarding doing a {{ 'full' if full_restart_cluster | bool else 'rolling'}} cluster restart."
  45. - debug: msg="Cluster logging-{{ _cluster_component }} was not in an optimal state and will not be automatically restarted. Please see documentation regarding doing a {{ 'full' if full_restart_cluster | bool else 'rolling'}} cluster restart."
  46. - when: _pod_status.stdout is undefined or ( (_pod_status.stdout | from_json)['status'] in ['green'] and (_pod_status.stdout | from_json)['number_of_nodes'] == _cluster_pods.stdout.split(' ') | count )
  47. block:
  48. - command: >
  49. {{ openshift_client_binary }}
  50. --config={{ openshift.common.config_base }}/master/admin.kubeconfig
  51. get dc
  52. -l component={{ _cluster_component }},provider=openshift
  53. -n {{ openshift_logging_elasticsearch_namespace }}
  54. -o jsonpath={.items[*].metadata.name}
  55. register: _cluster_dcs
  56. ## restart all dcs for full restart
  57. - name: "Performing full cluster restart for {{ _cluster_component }} cluster"
  58. include_tasks: full_cluster_restart.yml
  59. vars:
  60. logging_restart_cluster_dcs: "{{ _cluster_dcs.stdout_lines }}"
  61. when:
  62. - full_restart_cluster | bool
  63. ## restart the node if it's dc is in the list of nodes to restart
  64. - name: "Performing rolling cluster restart for {{ _cluster_component }} cluster"
  65. include_tasks: rolling_cluster_restart.yml
  66. vars:
  67. logging_restart_cluster_dcs: "{{ _restart_logging_nodes | intersect(_cluster_dcs.stdout) }}"
  68. when:
  69. - not full_restart_cluster | bool
  70. # remove temp dir
  71. - name: Cleaning up temp dir
  72. file:
  73. path: "{{ _logging_handler_tempdir.stdout }}"
  74. state: absent
  75. changed_when: False