123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341 |
- ---
- ###############################################################################
- # Upgrade Masters
- ###############################################################################
- # Prior to 3.6, openshift-ansible created etcd serving certificates
- # without a SubjectAlternativeName entry for the system hostname. The
- # SAN list in Go 1.8 is now (correctly) authoritative and since
- # openshift-ansible configures masters to talk to etcd hostnames
- # rather than IP addresses, we must correct etcd certificates.
- #
- # This play examines the etcd serving certificate SANs on each etcd
- # host and records whether or not the system hostname is missing.
- - name: Examine etcd serving certificate SAN
- hosts: oo_etcd_to_config
- tasks:
- - slurp:
- src: /etc/etcd/server.crt
- register: etcd_serving_cert
- - set_fact:
- __etcd_cert_lacks_hostname: "{{ (openshift.common.hostname not in (etcd_serving_cert.content | b64decode | lib_utils_oo_parse_certificate_san)) | bool }}"
- # Redeploy etcd certificates when hostnames were missing from etcd
- # serving certificate SANs.
- - import_playbook: ../../../openshift-etcd/redeploy-certificates.yml
- when:
- - true in hostvars | lib_utils_oo_select_keys(groups['oo_etcd_to_config']) | lib_utils_oo_collect('__etcd_cert_lacks_hostname') | default([false])
- - name: Backup and upgrade etcd
- import_playbook: ../../../openshift-etcd/private/upgrade_main.yml
- # Create service signer cert when missing. Service signer certificate
- # is added to master config in the master_config_upgrade hook.
- - name: Determine if service signer cert must be created
- hosts: oo_first_master
- tasks:
- - name: Determine if service signer certificate must be created
- stat:
- path: "{{ openshift.common.config_base }}/master/service-signer.crt"
- register: service_signer_cert_stat
- changed_when: false
- - import_playbook: create_service_signer_cert.yml
- # oc adm migrate storage should be run prior to etcd v3 upgrade
- # See: https://github.com/openshift/origin/pull/14625#issuecomment-308467060
- - name: Pre master upgrade - Upgrade all storage
- hosts: oo_first_master
- roles:
- - openshift_facts
- tasks:
- - name: Upgrade all storage
- command: >
- {{ openshift_client_binary }} adm --config={{ openshift.common.config_base }}/master/admin.kubeconfig
- migrate storage --include=* --confirm
- register: l_pb_upgrade_control_plane_pre_upgrade_storage
- when: openshift_upgrade_pre_storage_migration_enabled | default(true) | bool
- failed_when:
- - openshift_upgrade_pre_storage_migration_enabled | default(true) | bool
- - l_pb_upgrade_control_plane_pre_upgrade_storage.rc != 0
- - openshift_upgrade_pre_storage_migration_fatal | default(true) | bool
- # Set openshift_master_facts separately. In order to reconcile
- # admission_config's, we currently must run openshift_master_facts and
- # then run openshift_facts.
- - name: Set OpenShift master facts
- hosts: oo_masters_to_config
- roles:
- - openshift_master_facts
- # The main master upgrade play. Should handle all changes to the system in one pass, with
- # support for optional hooks to be defined.
- - name: Upgrade master
- hosts: oo_masters_to_config
- vars:
- openshift_master_ha: "{{ groups.oo_masters_to_config | length > 1 }}"
- serial: 1
- roles:
- - openshift_facts
- tasks:
- # Run the pre-upgrade hook if defined:
- - debug: msg="Running master pre-upgrade hook {{ openshift_master_upgrade_pre_hook }}"
- when: openshift_master_upgrade_pre_hook is defined
- - include_tasks: "{{ openshift_master_upgrade_pre_hook }}"
- when: openshift_master_upgrade_pre_hook is defined
- - import_role:
- name: openshift_master
- tasks_from: upgrade.yml
- # Run the upgrade hook prior to restarting services/system if defined:
- - debug: msg="Running master upgrade hook {{ openshift_master_upgrade_hook }}"
- when: openshift_master_upgrade_hook is defined
- - include_tasks: "{{ openshift_master_upgrade_hook }}"
- when: openshift_master_upgrade_hook is defined
- - include_tasks: ../../../openshift-master/private/tasks/restart_hosts.yml
- when: openshift.common.rolling_restart_mode == 'system'
- - include_tasks: ../../../openshift-master/private/tasks/restart_services.yml
- when: openshift.common.rolling_restart_mode == 'services'
- # Run the post-upgrade hook if defined:
- - debug: msg="Running master post-upgrade hook {{ openshift_master_upgrade_post_hook }}"
- when: openshift_master_upgrade_post_hook is defined
- - include_tasks: "{{ openshift_master_upgrade_post_hook }}"
- when: openshift_master_upgrade_post_hook is defined
- - name: Post master upgrade - Upgrade clusterpolicies storage
- command: >
- {{ openshift_client_binary }} adm --config={{ openshift.common.config_base }}/master/admin.kubeconfig
- migrate storage --include=clusterpolicies --confirm
- register: l_pb_upgrade_control_plane_post_upgrade_storage
- when:
- - openshift_upgrade_post_storage_migration_enabled | default(true) | bool
- - openshift_version is version_compare('3.7','<')
- failed_when:
- - openshift_upgrade_post_storage_migration_enabled | default(true) | bool
- - l_pb_upgrade_control_plane_post_upgrade_storage.rc != 0
- - openshift_upgrade_post_storage_migration_fatal | default(false) | bool
- run_once: true
- delegate_to: "{{ groups.oo_first_master.0 }}"
- - set_fact:
- master_update_complete: True
- ##############################################################################
- # Gate on master update complete
- ##############################################################################
- - name: Gate on master update
- hosts: localhost
- connection: local
- tasks:
- - set_fact:
- master_update_completed: "{{ hostvars
- | lib_utils_oo_select_keys(groups.oo_masters_to_config)
- | lib_utils_oo_collect('inventory_hostname', {'master_update_complete': true}) }}"
- - set_fact:
- master_update_failed: "{{ groups.oo_masters_to_config | difference(master_update_completed) | list }}"
- - fail:
- msg: "Upgrade cannot continue. The following masters did not finish updating: {{ master_update_failed | join(',') }}"
- when: master_update_failed | length > 0
- ###############################################################################
- # Reconcile Cluster Roles, Cluster Role Bindings and Security Context Constraints
- ###############################################################################
- - name: Reconcile Cluster Roles and Cluster Role Bindings and Security Context Constraints
- hosts: oo_masters_to_config
- roles:
- - { role: openshift_cli }
- - { role: openshift_facts }
- vars:
- __master_shared_resource_viewer_file: "shared_resource_viewer_role.yaml"
- tasks:
- - name: Reconcile Cluster Roles
- command: >
- {{ openshift_client_binary }} adm --config={{ openshift.common.config_base }}/master/admin.kubeconfig
- policy reconcile-cluster-roles --additive-only=true --confirm -o name
- register: reconcile_cluster_role_result
- when: openshift_version is version_compare('3.7','<')
- changed_when:
- - reconcile_cluster_role_result.stdout != ''
- - reconcile_cluster_role_result.rc == 0
- run_once: true
- - name: Reconcile Cluster Role Bindings
- command: >
- {{ openshift_client_binary }} adm --config={{ openshift.common.config_base }}/master/admin.kubeconfig
- policy reconcile-cluster-role-bindings
- --exclude-groups=system:authenticated
- --exclude-groups=system:authenticated:oauth
- --exclude-groups=system:unauthenticated
- --exclude-users=system:anonymous
- --additive-only=true --confirm -o name
- when: openshift_version is version_compare('3.7','<')
- register: reconcile_bindings_result
- changed_when:
- - reconcile_bindings_result.stdout != ''
- - reconcile_bindings_result.rc == 0
- run_once: true
- - name: Reconcile Jenkins Pipeline Role Bindings
- command: >
- {{ openshift_client_binary }} adm --config={{ openshift.common.config_base }}/master/admin.kubeconfig policy reconcile-cluster-role-bindings system:build-strategy-jenkinspipeline --confirm -o name
- run_once: true
- register: reconcile_jenkins_role_binding_result
- changed_when:
- - reconcile_jenkins_role_binding_result.stdout != ''
- - reconcile_jenkins_role_binding_result.rc == 0
- when:
- - openshift_version is version_compare('3.7','<')
- - when: openshift_upgrade_target is version_compare('3.7','<')
- block:
- - name: Retrieve shared-resource-viewer
- oc_obj:
- state: list
- kind: role
- name: "shared-resource-viewer"
- namespace: "openshift"
- register: objout
- - name: Determine if shared-resource-viewer is protected
- set_fact:
- __shared_resource_viewer_protected: true
- when:
- - "'results' in objout"
- - "'results' in objout['results']"
- - "'annotations' in objout['results']['results'][0]['metadata']"
- - "'openshift.io/reconcile-protect' in objout['results']['results'][0]['metadata']['annotations']"
- - "objout['results']['results'][0]['metadata']['annotations']['openshift.io/reconcile-protect'] == 'true'"
- - copy:
- src: "{{ item }}"
- dest: "/tmp/{{ item }}"
- with_items:
- - "{{ __master_shared_resource_viewer_file }}"
- when: __shared_resource_viewer_protected is not defined
- - name: Fixup shared-resource-viewer role
- oc_obj:
- state: present
- kind: role
- name: "shared-resource-viewer"
- namespace: "openshift"
- files:
- - "/tmp/{{ __master_shared_resource_viewer_file }}"
- delete_after: true
- when: __shared_resource_viewer_protected is not defined
- register: result
- retries: 3
- delay: 5
- until: result.rc == 0
- ignore_errors: true
- - name: Reconcile Security Context Constraints
- command: >
- {{ openshift_client_binary }} adm policy --config={{ openshift.common.config_base }}/master/admin.kubeconfig reconcile-sccs --confirm --additive-only=true -o name
- register: reconcile_scc_result
- changed_when:
- - reconcile_scc_result.stdout != ''
- - reconcile_scc_result.rc == 0
- run_once: true
- - name: Migrate storage post policy reconciliation
- command: >
- {{ openshift_client_binary }} adm --config={{ openshift.common.config_base }}/master/admin.kubeconfig
- migrate storage --include=* --confirm
- run_once: true
- register: l_pb_upgrade_control_plane_post_upgrade_storage
- when: openshift_upgrade_post_storage_migration_enabled | default(true) | bool
- failed_when:
- - openshift_upgrade_post_storage_migration_enabled | default(true) | bool
- - l_pb_upgrade_control_plane_post_upgrade_storage.rc != 0
- - openshift_upgrade_post_storage_migration_fatal | default(false) | bool
- - set_fact:
- reconcile_complete: True
- ##############################################################################
- # Gate on reconcile
- ##############################################################################
- - name: Gate on reconcile
- hosts: localhost
- connection: local
- tasks:
- - set_fact:
- reconcile_completed: "{{ hostvars
- | lib_utils_oo_select_keys(groups.oo_masters_to_config)
- | lib_utils_oo_collect('inventory_hostname', {'reconcile_complete': true}) }}"
- - set_fact:
- reconcile_failed: "{{ groups.oo_masters_to_config | difference(reconcile_completed) | list }}"
- - fail:
- msg: "Upgrade cannot continue. The following masters did not finish reconciling: {{ reconcile_failed | join(',') }}"
- when: reconcile_failed | length > 0
- - name: Upgrade Docker on dedicated containerized etcd hosts
- hosts: oo_etcd_to_config:!oo_nodes_to_upgrade
- serial: 1
- any_errors_fatal: true
- roles:
- - openshift_facts
- tasks:
- - include_tasks: docker/tasks/upgrade.yml
- when: l_docker_upgrade is defined and l_docker_upgrade | bool and not openshift_is_atomic | bool
- - name: Drain and upgrade master nodes
- hosts: oo_masters_to_config:&oo_nodes_to_upgrade
- # This var must be set with -e on invocation, as it is not a per-host inventory var
- # and is evaluated early. Values such as "20%" can also be used.
- serial: "{{ openshift_upgrade_control_plane_nodes_serial | default(1) }}"
- max_fail_percentage: "{{ openshift_upgrade_control_plane_nodes_max_fail_percentage | default(0) }}"
- pre_tasks:
- - name: Load lib_openshift modules
- import_role:
- name: lib_openshift
- # TODO: To better handle re-trying failed upgrades, it would be nice to check if the node
- # or docker actually needs an upgrade before proceeding. Perhaps best to save this until
- # we merge upgrade functionality into the base roles and a normal config.yml playbook run.
- - name: Mark node unschedulable
- oc_adm_manage_node:
- node: "{{ openshift.node.nodename | lower }}"
- schedulable: False
- delegate_to: "{{ groups.oo_first_master.0 }}"
- retries: 10
- delay: 5
- register: node_unschedulable
- until: node_unschedulable is succeeded
- - name: Drain Node for Kubelet upgrade
- command: >
- {{ hostvars[groups.oo_first_master.0]['first_master_client_binary'] }} adm drain {{ openshift.node.nodename | lower }}
- --config={{ openshift.common.config_base }}/master/admin.kubeconfig
- --force --delete-local-data --ignore-daemonsets
- --timeout={{ openshift_upgrade_nodes_drain_timeout | default(0) }}s
- delegate_to: "{{ groups.oo_first_master.0 }}"
- register: l_upgrade_control_plane_drain_result
- until: not (l_upgrade_control_plane_drain_result is failed)
- retries: "{{ 1 if ( openshift_upgrade_nodes_drain_timeout | default(0) | int ) == 0 else 0 }}"
- delay: 5
- failed_when:
- - l_upgrade_control_plane_drain_result is failed
- - openshift_upgrade_nodes_drain_timeout | default(0) | int == 0
- roles:
- - openshift_facts
- post_tasks:
- - import_role:
- name: openshift_node
- tasks_from: upgrade.yml
- - import_role:
- name: openshift_manage_node
- tasks_from: config.yml
- vars:
- openshift_master_host: "{{ groups.oo_first_master.0 }}"
|