|
@@ -1,37 +1,5 @@
|
|
|
---
|
|
|
###############################################################################
|
|
|
-# The restart playbook should be run after this playbook completes.
|
|
|
-###############################################################################
|
|
|
-
|
|
|
-# Separate step so we can execute in parallel and clear out anything unused
|
|
|
-# before we get into the serialized upgrade process which will then remove
|
|
|
-# remaining images if possible.
|
|
|
-- name: Cleanup unused Docker images
|
|
|
- hosts: oo_masters_to_config:oo_nodes_to_config:oo_etcd_to_config
|
|
|
- tasks:
|
|
|
- - name: Check Docker image count
|
|
|
- shell: "docker images -aq | wc -l"
|
|
|
- register: docker_image_count
|
|
|
- when: docker_upgrade_nuke_images is defined and docker_upgrade_nuke_images | bool
|
|
|
-
|
|
|
- - debug: var=docker_image_count.stdout
|
|
|
- when: docker_upgrade_nuke_images is defined and docker_upgrade_nuke_images | bool
|
|
|
-
|
|
|
- - name: Remove unused Docker images for Docker 1.10+ migration
|
|
|
- shell: "docker rmi `docker images -aq`"
|
|
|
- # Will fail on images still in use:
|
|
|
- failed_when: false
|
|
|
- when: docker_upgrade_nuke_images is defined and docker_upgrade_nuke_images | bool
|
|
|
-
|
|
|
- - name: Check Docker image count
|
|
|
- shell: "docker images -aq | wc -l"
|
|
|
- register: docker_image_count
|
|
|
- when: docker_upgrade_nuke_images is defined and docker_upgrade_nuke_images | bool
|
|
|
-
|
|
|
- - debug: var=docker_image_count.stdout
|
|
|
- when: docker_upgrade_nuke_images is defined and docker_upgrade_nuke_images | bool
|
|
|
-
|
|
|
-###############################################################################
|
|
|
# Upgrade Masters
|
|
|
###############################################################################
|
|
|
- name: Upgrade master packages
|
|
@@ -195,68 +163,12 @@
|
|
|
msg: "Upgrade cannot continue. The following masters did not finish reconciling: {{ reconcile_failed | join(',') }}"
|
|
|
when: reconcile_failed | length > 0
|
|
|
|
|
|
-###############################################################################
|
|
|
-# Upgrade Nodes
|
|
|
-###############################################################################
|
|
|
-
|
|
|
-# Here we handle all tasks that might require a node evac. (upgrading docker, and the node service)
|
|
|
-- name: Perform upgrades that may require node evacuation
|
|
|
- hosts: oo_masters_to_config:oo_etcd_to_config:oo_nodes_to_config
|
|
|
+- name: Upgrade Docker on dedicated containerized etcd hosts
|
|
|
+ hosts: oo_etcd_to_config:!oo_nodes_to_config
|
|
|
serial: 1
|
|
|
any_errors_fatal: true
|
|
|
roles:
|
|
|
- openshift_facts
|
|
|
- handlers:
|
|
|
- - include: ../../../../roles/openshift_node/handlers/main.yml
|
|
|
- static: yes
|
|
|
tasks:
|
|
|
- # TODO: To better handle re-trying failed upgrades, it would be nice to check if the node
|
|
|
- # or docker actually needs an upgrade before proceeding. Perhaps best to save this until
|
|
|
- # we merge upgrade functionality into the base roles and a normal config.yml playbook run.
|
|
|
- - name: Determine if node is currently scheduleable
|
|
|
- command: >
|
|
|
- {{ openshift.common.client_binary }} get node {{ openshift.node.nodename }} -o json
|
|
|
- register: node_output
|
|
|
- delegate_to: "{{ groups.oo_first_master.0 }}"
|
|
|
- changed_when: false
|
|
|
- when: inventory_hostname in groups.oo_nodes_to_config
|
|
|
-
|
|
|
- - set_fact:
|
|
|
- was_schedulable: "{{ 'unschedulable' not in (node_output.stdout | from_json).spec }}"
|
|
|
- when: inventory_hostname in groups.oo_nodes_to_config
|
|
|
-
|
|
|
- - name: Mark unschedulable if host is a node
|
|
|
- command: >
|
|
|
- {{ openshift.common.admin_binary }} manage-node {{ openshift.node.nodename }} --schedulable=false
|
|
|
- delegate_to: "{{ groups.oo_first_master.0 }}"
|
|
|
- when: inventory_hostname in groups.oo_nodes_to_config
|
|
|
-
|
|
|
- - name: Evacuate Node for Kubelet upgrade
|
|
|
- command: >
|
|
|
- {{ openshift.common.admin_binary }} manage-node {{ openshift.node.nodename }} --evacuate --force
|
|
|
- delegate_to: "{{ groups.oo_first_master.0 }}"
|
|
|
- when: inventory_hostname in groups.oo_nodes_to_config
|
|
|
-
|
|
|
- include: docker/upgrade.yml
|
|
|
when: l_docker_upgrade is defined and l_docker_upgrade | bool and not openshift.common.is_atomic | bool
|
|
|
- - include: "{{ node_config_hook }}"
|
|
|
- when: node_config_hook is defined and inventory_hostname in groups.oo_nodes_to_config
|
|
|
-
|
|
|
- - include: rpm_upgrade.yml
|
|
|
- vars:
|
|
|
- component: "node"
|
|
|
- openshift_version: "{{ openshift_pkg_version | default('') }}"
|
|
|
- when: inventory_hostname in groups.oo_nodes_to_config and not openshift.common.is_containerized | bool
|
|
|
-
|
|
|
- - include: containerized_node_upgrade.yml
|
|
|
- when: inventory_hostname in groups.oo_nodes_to_config and openshift.common.is_containerized | bool
|
|
|
-
|
|
|
- - meta: flush_handlers
|
|
|
-
|
|
|
- - name: Set node schedulability
|
|
|
- command: >
|
|
|
- {{ openshift.common.admin_binary }} manage-node {{ openshift.node.nodename }} --schedulable=true
|
|
|
- delegate_to: "{{ groups.oo_first_master.0 }}"
|
|
|
- when: inventory_hostname in groups.oo_nodes_to_config and was_schedulable | bool
|
|
|
-
|
|
|
-
|